Feature Selection based Classification using Naive Bayes, J48 and Support Vector Machine

  • Bhosale D
  • Ade R
  • R. Deshmukh P
N/ACitations
Citations of this article
22Readers
Mendeley users who have this article in their library.

Abstract

One way to improve accuracy of a classifier is to use the minimum number of features. Many feature selection techniques are proposed to find out the most important features. In this paper, feature selection methods Co-relation based feature Selection, Wrapper method and Information Gain are used, before applying supervised learning based classification techniques. The results show that Support vector Machine with Information Gain and Wrapper method have the best results as compared to others tested.

Cite

CITATION STYLE

APA

Bhosale, D., Ade, R., & R. Deshmukh, P. (2014). Feature Selection based Classification using Naive Bayes, J48 and Support Vector Machine. International Journal of Computer Applications, 99(16), 14–18. https://doi.org/10.5120/17456-8202

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free