We introduce a novel, robust data-driven regularization strategy called Adaptive Regularized Boosting (AR-Boost), motivated by a desire to reduce overfitting. We replace AdaBoost's hard margin with a regularized soft margin that trades-off between a larger margin, at the expense of misclassification errors. Minimizing this regularized exponential loss results in a boosting algorithm that relaxes the weak learning assumption further: it can use classifiers with error greater than 1/2. This enables a natural extension to multiclass boosting, and further reduces overfitting in both the binary and multiclass cases. We derive bounds for training and generalization errors, and relate them to AdaBoost. Finally, we show empirical results on benchmark data that establish the robustness of our approach and improved performance overall. © 2013 Springer-Verlag.
CITATION STYLE
Saha, B. N., Kunapuli, G., Ray, N., Maldjian, J. A., & Natarajan, S. (2013). AR-Boost: Reducing overfitting by a robust data-driven regularization strategy. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8190 LNAI, pp. 1–16). https://doi.org/10.1007/978-3-642-40994-3_1
Mendeley helps you to discover research relevant for your work.