Tree-based ensembles with dynamic soft feature selection

7Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Tree-based ensembles have been proven to be among the most accurate and versatile state-of-the-art learning machines. The best known are MART (gradient tree boosting) and RF (Random Forest.) Usage of such ensembles in supervised problems with a very high dimensional input space can be challenging. Modelling with MART becomes computationally infeasible, and RF can produce low quality models when only a small subset of predictors is relevant. We propose an importance based sampling scheme where only a small sample of variables is selected at every step of ensemble construction. The sampling distribution is modified at every iteration to promote variables more relevant to the target. Experiments show that this method gives MART a very substantial performance boost with at least the same level of accuracy. It also adds a bias correction element to RF for very noisy problems. MART with dynamic feature selection produced very competitive results at the NIPS-2003 feature selection challenge. © Springer-Verlag Berlin Heidelberg 2006.

Cite

CITATION STYLE

APA

Borisov, A., Eruhimov, V., & Tuv, E. (2006). Tree-based ensembles with dynamic soft feature selection. Studies in Fuzziness and Soft Computing, 207, 359–374. https://doi.org/10.1007/978-3-540-35488-8_16

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free