Single estimator versus bagging
  • References/Python/scikit-learn/Examples/Ensemble methods

This example illustrates and compares the bias-variance decomposition of the expected mean squared error of a single estimator against

2025-01-10 15:47:30
Two-class AdaBoost
  • References/Python/scikit-learn/Examples/Ensemble methods

This example fits an AdaBoosted decision stump on a non-linearly separable classification dataset composed of two ?Gaussian quantiles? clusters (see

2025-01-10 15:47:30
Plot class probabilities calculated by the VotingClassifier
  • References/Python/scikit-learn/Examples/Ensemble methods

Plot the class probabilities of the first sample in a toy dataset predicted by three different classifiers and averaged by the

2025-01-10 15:47:30
Gradient Boosting regularization
  • References/Python/scikit-learn/Examples/Ensemble methods

Illustration of the effect of different regularization strategies for Gradient Boosting. The example is taken from Hastie et al 2009. The loss function

2025-01-10 15:47:30
Multi-class AdaBoosted Decision Trees
  • References/Python/scikit-learn/Examples/Ensemble methods

This example reproduces Figure 1 of Zhu et al [1] and shows how boosting can improve prediction accuracy on a multi-class problem. The classification dataset

2025-01-10 15:47:30
Decision Tree Regression with AdaBoost
  • References/Python/scikit-learn/Examples/Ensemble methods

A decision tree is boosted using the AdaBoost.R2 [1] algorithm on a 1D sinusoidal dataset with a small amount of Gaussian noise. 299 boosts (300 decision

2025-01-10 15:47:30
Feature transformations with ensembles of trees
  • References/Python/scikit-learn/Examples/Ensemble methods

Transform your features into a higher dimensional, sparse space. Then train a linear model on these features. First fit an ensemble of

2025-01-10 15:47:30
Gradient Boosting regression
  • References/Python/scikit-learn/Examples/Ensemble methods

Demonstrate Gradient Boosting on the Boston housing dataset. This example fits a Gradient Boosting model with least squares loss and 500 regression trees

2025-01-10 15:47:30
Plot the decision surfaces of ensembles of trees on the iris dataset
  • References/Python/scikit-learn/Examples/Ensemble methods

Plot the decision surfaces of forests of randomized trees trained on pairs of features of the iris dataset. This

2025-01-10 15:47:30
Plot the decision boundaries of a VotingClassifier
  • References/Python/scikit-learn/Examples/Ensemble methods

Plot the decision boundaries of a VotingClassifier for two features of the Iris dataset. Plot the class probabilities

2025-01-10 15:47:30