Measuring the stability of feature selection with applications to ensemble methods

23Citations
Citations of this article
26Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Ensemble methods are often used to decide on a good selection of features for later processing by a classifier. Examples of this are in the determination of Random Forest variable importance proposed by Breiman, and in the concept of feature selection ensembles, where the outputs of multiple feature selectors are combined to yield more robust results. All of these methods rely critically on the concept of feature selection stability - similar but distinct to the concept of diversity in classifier ensembles. We conduct a systematic study of the literature, identifying desirable/undesirable properties, and identify a weakness in existing measures. A simple correction is proposed, and empirical studies are conducted to illustrate its utility.

Author supplied keywords

Cite

CITATION STYLE

APA

Nogueira, S., & Brown, G. (2015). Measuring the stability of feature selection with applications to ensemble methods. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9132, pp. 135–146). Springer Verlag. https://doi.org/10.1007/978-3-319-20248-8_12

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free