Combining naive Bayes and n-gram language models for text classification

106Citations
Citations of this article
131Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We augment the naive Bayes model with an n-gram language model to address two shortcomings of naive Bayes text classifiers. The chain augmented naive Bayes classifiers we propose have two advantages over standard naive Bayes classifiers. First, a chain augmented naive Bayes model relaxes some of the independence assumptions of naive Bayes - allowing a local Markov chain dependence in the observed variables - while still permitting efficient inference and learning. Second, smoothing techniques from statistical language modeling can be used to recover better estimates than the Laplace smoothing techniques usually used in naive Bayes classification. Our experimental results on three real world data sets show that we achieve substantial improvements over standard naive Bayes classification, while also achieving state of the art performance that competes with the best known methods in these cases. © Springer-Verlag Berlin Heidelberg 2003.

Cite

CITATION STYLE

APA

Peng, F., & Schuurmans, D. (2003). Combining naive Bayes and n-gram language models for text classification. Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2633, 335–350. https://doi.org/10.1007/3-540-36618-0_24

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free