Skip to content

Boosting

  • Trains multiple weak learners sequentially so each new learner focuses on the mistakes of the previous ones.
  • The final predictor is a weighted combination of the weak learners; weights reflect each learner’s performance.
  • Commonly improves accuracy on small or imbalanced datasets and can model non-linear relationships.

Boosting is a machine learning ensemble method that combines multiple weak learners to create a stronger model. A weak learner is a model that performs slightly better than random guessing. The boosting algorithm trains the weak learners sequentially, with each subsequent learner trying to correct the mistakes of the previous one. The final model is the weighted combination of all the weak learners, where the weights are determined by the performance of each learner.

  • Training procedure: Weak learners are fitted one after another. After each learner makes predictions on the training set, the algorithm increases the importance (weights) of the data points that were misclassified so that subsequent learners focus more on those errors.
  • Final model: Predictions from all weak learners are combined into a single model by weighting each learner according to its performance.
  • Notable algorithms: AdaBoost (Adaptive Boosting) introduced this reweighting idea in 1996. Gradient-boosting variants include XGBoost (eXtreme Gradient Boosting) and LightGBM (Light Gradient Boosting Machine). XGBoost exposes hyperparameters such as learning rate, number of trees, and maximum tree depth. LightGBM uses histogram-based algorithms and decision tree learning to improve training efficiency and is designed for large-scale data and distributed training.
  • Advantages cited in the source: can achieve better performance with less training data, can reduce overfitting (because weak learners are trained on different subsets and the final model combines them), handles imbalanced datasets by adjusting example weights, and can capture non-linear relationships by combining multiple weak learners.
  • Dataset: 100 points, 50 positive (labeled 1) and 50 negative (labeled 0).
  • Initial weights: each data point has weight 1/100.
  • After training the first weak learner: it correctly classifies 45 positive and 45 negative points, misclassifying 5 positive and 5 negative points (10 misclassified total). The algorithm increases the weights of those 10 misclassified points so the next learner focuses more on them.
  • Iteration and weighting: the process repeats for a predetermined number of iterations. The final model is a weighted combination of learners; for example, a first learner that correctly classified 45 positive and 45 negative points would receive a higher weight than a second learner that correctly classified 44 positive and 44 negative points.
  • Use case: classify emails as spam or not spam.
  • Procedure: train multiple weak learners (e.g., decision trees or logistic regression) sequentially, adjusting data-point weights based on prediction accuracy. The final model combines learners with weights determined by their performance.
  • Benefit: can handle imbalanced classes by increasing weights on minority-class examples.
  • Use case: predict a continuous target such as future stock price using historical prices.
  • Procedure: train multiple weak learners (e.g., linear regression or decision trees) sequentially, reweighting data points and combining learners into a final weighted model.
  • Benefit: can model non-linear relationships between features and the target.
  • Binary classification: spam detection, credit default prediction.
  • Regression: stock price prediction.
  • Industry applications mentioned: online advertising, recommendation systems, credit risk analysis.
  • Boosting can achieve high accuracy with less training data by focusing subsequent learners on previous errors.
  • It can reduce overfitting because weak learners are trained on different subsets and the final model aggregates them.
  • Boosting algorithms can handle imbalanced datasets by adjusting example weights and can learn non-linear relationships through the combination of multiple weak learners.
  • AdaBoost (Adaptive Boosting)
  • Gradient boosting
  • XGBoost (eXtreme Gradient Boosting)
  • LightGBM (Light Gradient Boosting Machine)
  • Weak learner
  • Decision tree
  • Logistic regression
  • Linear regression