Unified Robust Boosting
Pub. online: 28 June 2024
Type: Computing In Data Science
Open Access
Received
18 March 2024
18 March 2024
Accepted
22 April 2024
22 April 2024
Published
28 June 2024
28 June 2024
Abstract
Boosting is a popular algorithm in supervised machine learning with wide applications in regression and classification problems. It combines weak learners, such as regression trees, to obtain accurate predictions. However, in the presence of outliers, traditional boosting may yield inferior results since the algorithm optimizes a convex loss function. Recent literature has proposed boosting algorithms that optimize robust nonconvex loss functions. Nevertheless, there is a lack of weighted estimation to indicate the outlier status of observations. This article introduces the iteratively reweighted boosting (IRBoost) algorithm, which combines robust loss optimization and weighted estimation. It can be conveniently constructed with existing software. The output includes weights as valuable diagnostics for the outlier status of observations. For practitioners interested in the boosting algorithm, the new method can be interpreted as a way to tune robust observation weights. IRBoost is implemented in the R package irboost and is demonstrated using publicly available data in generalized linear models, classification, and survival data analysis.
Supplementary material
Supplementary MaterialThe R code necessary to reproduce the analysis presented in the manuscript is provided.
References
Barnwal A, Cho H, Hocking T (2022). Survival regression with accelerated failure time model in XGBoost. Journal of Computational and Graphical Statistics, 31(4): 1292–1302. https://doi.org/10.1080/10618600.2022.2067548
Friedman J (2001). Greedy function approximation: A gradient boosting machine. The Annals of Statistics, 29(5): 1189–1232. https://doi.org/10.1214/aos/1013203451
Friedman J, Hastie T, Tibshirani R (2000). Additive logistic regression: A statistical view of boosting (with discussion and a rejoinder by the authors). The Annals of Statistics, 28(2): 337–407. https://doi.org/10.1214/aos/1016218223
Long PM, Servedio RA (2010). Random classification noise defeats all convex potential boosters. Machine Learning, 78(3): 287–304. https://doi.org/10.1007/s10994-009-5165-z
Park SY, Liu Y (2011). Robust penalized logistic regression with truncated loss functions. Canadian Journal of Statistics, 39(2): 300–323. https://doi.org/10.1002/cjs.10105
Sigrist F (2021). Gradient and Newton boosting for classification and regression. Expert Systems with Applications, 167: 114080. https://doi.org/10.1016/j.eswa.2020.114080
Wang Z (2018a). Quadratic majorization for nonconvex loss with applications to the boosting algorithm. Journal of Computational and Graphical Statistics, 27(3): 491–502. https://doi.org/10.1080/10618600.2018.1424635
Wang Z (2018b). Robust boosting with truncated loss functions. Electronic Journal of Statistics, 12(1): 599–650. https://doi.org/10.1214/18-EJS1434
Wang Z (2024b). Unified robust estimation. Australian & New Zealand Journal of Statistics, 66(1): 77–102. https://doi.org/10.1111/anzs.12409
Wu Y, Liu Y (2007). Robust truncated hinge loss support vector machines. Journal of the American Statistical Association, 102(479): 974–983. https://doi.org/10.1198/016214507000000617