Searching parsimonious solutions with GA-PARSIMONY and XGboost in high-dimensional databases

15Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.
Get full text

Abstract

EXtreme Gradient Boosting (XGBoost) has become one of the most successful techniques in machine learning competitions. It is computationally efficient and scalable, it supports a wide variety of objective functions and it includes different mechanisms to avoid overfitting and improve accuracy. Having so many tuning parameters, soft computing (SC) is an alternative to search precise and robust models against classical hyper-tuning methods. In this context, we present a preliminary study in which a SC methodology, named GA-PARSIMONY, is used to find accurate and parsimonious XGBoost solutions. The methodology was designed to optimize the search of parsimonious models by feature selection, parameter tuning and model selection. In this work, different experiments are conducted with four complexity metrics in six high dimensional datasets. Although XGBoost performs well with high-dimensional databases, preliminary results indicated that GAPARSIMONY with feature selection slightly improved the testing error. Therefore, the choice of solutions with fewer inputs, between those with similar cross-validation errors, can help to obtain more robust solutions with better generalization capabilities.

Cite

CITATION STYLE

APA

Martinez-de-Pison, F. J., Fraile-Garcia, E., Ferreiro-Cabello, J., Gonzalez, R., & Pernia, A. (2017). Searching parsimonious solutions with GA-PARSIMONY and XGboost in high-dimensional databases. In Advances in Intelligent Systems and Computing (Vol. 527, pp. 201–210). Springer Verlag. https://doi.org/10.1007/978-3-319-47364-2_20

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free