You are on page 1of 2

Ensemble learning refers to algorithms that combine the predictions from two or more models.

 Bagging involves fitting many decision trees on different samples of the same dataset and
averaging the predictions.
 Stacking involves fitting many different models types on the same data and using another
model to learn how to best combine the predictions.
 Boosting involves adding ensemble members sequentially that correct the predictions
made by prior models and outputs a weighted average of the predictions.
Random Forest is a classifier that contains a number of decision trees on
various subsets of the given dataset and takes the average to improve the
predictive accuracy of that dataset." Instead of relying on one decision tree, the
random forest takes the prediction from each tree and based on the majority votes
of predictions, and it predicts the final output.

You might also like