Bayesian model averaging naive bayes (BMA-NB): Averaging over an exponential number of feature models in linear time

4Citations
Citations of this article
23Readers
Mendeley users who have this article in their library.

Abstract

Naive Bayes (NB) is well-known to be a simple but effective classifier, especially when combined with feature selection. Unfortunately, feature selection methods are often greedy and thus cannot guarantee an optimal feature set is selected. An alternative to feature selection is to use Bayesian model averaging (BMA), which computes a weighted average over multiple predictors; when the different predictor models correspond to different feature sets, BMA has the advantage over feature selection that its predictions tend to have lower variance on average in comparison to any single model. In this paper, we show for the first time that it is possible to exactly evaluate BMA over the exponentiallysized powerset of NB feature models in linear-time in the number of features; this yields an algorithm about as expensive to train as a single NB model with all features, but yet provably converges to the globally optimal feature subset in the asymptotic limit of data. We evaluate this novel BMA-NB classifier on a range of datasets showing that it never underperforms NB (as expected) and sometimes offers performance competitive (or superior) to classifiers such as SVMs and logistic regression while taking a fraction of the time to train.

Cite

CITATION STYLE

APA

Wu, G., Sanner, S., & Oliveira, R. F. S. C. (2015). Bayesian model averaging naive bayes (BMA-NB): Averaging over an exponential number of feature models in linear time. In Proceedings of the National Conference on Artificial Intelligence (Vol. 4, pp. 3094–3100). AI Access Foundation. https://doi.org/10.1609/aaai.v29i1.9634

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free