An approach to evolutionary ensemble learning for classification is proposed using genetic programming in which boosting is used to construct a stack of programs. Each application of boosting identifies a single champion and a residual dataset, i.e. the training records that thus far were not correctly classified. The next program is only trained against the residual, with the process iterating until some maximum ensemble size or no further residual remains. Training against a residual dataset actively reduces the cost of training. Deploying the ensemble as a stack also means that only one classifier might be necessary to make a prediction, so improving interpretability. Benchmarking studies are conducted to illustrate competitiveness with the prediction accuracy of current state-of-the-art evolutionary ensemble learning algorithms, while providing solutions that are orders of magnitude simpler. Further benchmarking with a high cardinality dataset indicates that the proposed method is also more accurate and efficient than XGBoost.
CITATION STYLE
Zhou, Z., Qiu, Z., Niblett, B., Johnston, A., Schwartzentruber, J., Zincir-Heywood, N., & Heywood, M. I. (2023). A Boosting Approach to Constructing an Ensemble Stack. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 13986 LNCS, pp. 133–148). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-031-29573-7_9
Mendeley helps you to discover research relevant for your work.