A comparison of model aggregation methods for regression

17Citations
Citations of this article
28Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Combining machine learning models is a means of improving overall accuracy. Various algorithms have been proposed to create aggregate models from other models, and two popular examples for classification are Bagging and AdaBoost. In this paper we examine their adaptation to regression, and benchmark them on synthetic and real-world data. Our experiments reveal that different types of AdaBoost algorithms require different complexities of base models. They out-perform Bagging at their best, but Bagging achieves a consistent level of success with all base models, providing a robust alternative. © Springer-Verlag Berlin Heidelberg 2003.

Cite

CITATION STYLE

APA

Barutçuoǧlu, Z., & Alpaydin, E. (2003). A comparison of model aggregation methods for regression. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). https://doi.org/10.1007/3-540-44989-2_10

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free