Search a title or topic

Over 20 million podcasts, powered by 

Player FM logo

Content provided by Richard M. Golden, M.S.E.E., and B.S.E.E.. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Richard M. Golden, M.S.E.E., and B.S.E.E. or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://staging.podcastplayer.com/legal.
Player FM - Podcast App
Go offline with the Player FM app!

LM101-077: How to Choose the Best Model using BIC

24:15
 
Share
 

Manage episode 232660961 series 2497400
Content provided by Richard M. Golden, M.S.E.E., and B.S.E.E.. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Richard M. Golden, M.S.E.E., and B.S.E.E. or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://staging.podcastplayer.com/legal.

In this 77th episode of www.learningmachines101.com , we explain the proper semantic interpretation of the Bayesian Information Criterion (BIC) and emphasize how this semantic interpretation is fundamentally different from AIC (Akaike Information Criterion) model selection methods. Briefly, BIC is used to estimate the probability of the training data given the probability model, while AIC is used to estimate out-of-sample prediction error. The probability of the training data given the model is called the “marginal likelihood”. Using the marginal likelihood, one can calculate the probability of a model given the training data and then use this analysis to support selecting the most probable model, selecting a model that minimizes expected risk, and support Bayesian model averaging. The assumptions which are required for BIC to be a valid approximation for the probability of the training data given the probability model are also discussed.

  continue reading

85 episodes

iconShare
 
Manage episode 232660961 series 2497400
Content provided by Richard M. Golden, M.S.E.E., and B.S.E.E.. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Richard M. Golden, M.S.E.E., and B.S.E.E. or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://staging.podcastplayer.com/legal.

In this 77th episode of www.learningmachines101.com , we explain the proper semantic interpretation of the Bayesian Information Criterion (BIC) and emphasize how this semantic interpretation is fundamentally different from AIC (Akaike Information Criterion) model selection methods. Briefly, BIC is used to estimate the probability of the training data given the probability model, while AIC is used to estimate out-of-sample prediction error. The probability of the training data given the model is called the “marginal likelihood”. Using the marginal likelihood, one can calculate the probability of a model given the training data and then use this analysis to support selecting the most probable model, selecting a model that minimizes expected risk, and support Bayesian model averaging. The assumptions which are required for BIC to be a valid approximation for the probability of the training data given the probability model are also discussed.

  continue reading

85 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Listen to this show while you explore
Play