Tree-based ensembles have been proven to be among the most accurate and versatile state-of-the-art learning machines. The best known are MART (gradient tree boosting) and RF (Random Forest.) Usage of such ensembles in supervised problems with a very high dimensional input space can be challenging. Modelling with MART becomes computationally infeasible, and RF can produce low quality models when only a small subset of predictors is relevant. We propose an importance based sampling scheme where only a small sample of variables is selected at every step of ensemble construction. The sampling distribution is modified at every iteration to promote variables more relevant to the target. Experiments show that this method gives MART a very substantial performance boost with at least the same level of accuracy. It also adds a bias correction element to RF for very noisy problems. MART with dynamic feature selection produced very competitive results at the NIPS-2003 feature selection challenge. © Springer-Verlag Berlin Heidelberg 2006.
CITATION STYLE
Borisov, A., Eruhimov, V., & Tuv, E. (2006). Tree-based ensembles with dynamic soft feature selection. Studies in Fuzziness and Soft Computing, 207, 359–374. https://doi.org/10.1007/978-3-540-35488-8_16
Mendeley helps you to discover research relevant for your work.