Obtaining Calibrated Probabilities from Boosting

Boosted decision trees typically yield good accuracy, precision, and ROC area. However, because the outputs from boosting are not well calibrated posterior probabilities, boosting yields poor squared error and cross-entropy. We empirically demonstrate why AdaBoost predicts distorted probabilities and examine three calibration methods for correcting this distortion: Platt Scaling, Isotonic Regression, and Logistic Correction. We also experiment with boosting using log-loss instead of the usual exponential loss. Experiments show that Logistic Correction and boosting with log-loss work well when boosting weak models such as decision stumps, but yield poor performance when boosting more complex models such as full decision trees. Platt Scaling and Isotonic Regression, however, significantly improve the probabilities predicted by

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/08/2019

Provably Robust Boosted Decision Stumps and Trees against Adversarial Attacks

The problem of adversarial samples has been studied extensively for neur...
research
06/08/2023

Boosting with Tempered Exponential Measures

One of the most popular ML algorithms, AdaBoost, can be derived from the...
research
10/16/2017

Calibrated Boosting-Forest

Excellent ranking power along with well calibrated probability estimates...
research
02/21/2023

Variational Boosted Soft Trees

Gradient boosting machines (GBMs) based on decision trees consistently d...
research
09/05/2011

Learning Nonlinear Functions Using Regularized Greedy Forest

We consider the problem of learning a forest of nonlinear decision rules...
research
07/31/2018

Probability Calibration Trees

Obtaining accurate and well calibrated probability estimates from classi...
research
10/29/2019

Minimal Variance Sampling in Stochastic Gradient Boosting

Stochastic Gradient Boosting (SGB) is a widely used approach to regulari...

Please sign up or login with your details

Forgot password? Click here to reset