Machine Learning Bytes

Bagging

07.24.2019 - By Erik PartridgePlay

Download our free app to listen on your phone

Download on the App StoreGet it on Google Play

Bagging is an ensemble meta-algorithm. Basically, we take some number of estimators (usually dozens-ish), train them each on some random subset of the training data. Then, we average the predictions of each individual estimator in order to make the resulting prediction. While this reduces the variance of your predictions (indeed, that is the core purpose of bagging), it may come at the trade off of bias.

For a more academic basis, see slide #13 of this lecture by Joëlle Pineau at McGill University.

---

Send in a voice message: https://podcasters.spotify.com/pod/show/mlbytes/message

More episodes from Machine Learning Bytes