07.24.2019 - By Machine Learning Bytes
Bagging is an ensemble meta-algorithm. Basically, we take some number of estimators (usually dozens-ish), train them each on some random subset of the training data. Then, we average the predictions of each individual estimator in order to make the resulting prediction. While this reduces the variance of your predictions (indeed, that is the core purpose of bagging), it may come at the trade off of bias.
For a more academic basis, see slide #13 of this lecture by Joëlle Pineau at McGill University.
Send in a voice message: https://anchor.fm/mlbytes/message