Using negation and phrases in inducing rules for text classification

0Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

An investigation into the use of negation in Inductive Rule Learning (IRL) for text classification is described. The use of negated features in the IRL process has been shown to improve effectiveness of classification. However, although in the case of small datasets it is perfectly feasible to include the potential negation of all possible features as part of the feature space, this is not possible for datasets that include large numbers of features such as those used in text mining applications. Instead a process whereby features to be negated can be identified dynamically is required. Such a process is described in the paper and compared with established techniques (JRip, NaiveBayes, Sequential Minimal Optimization (SMO), OlexGreedy). The work is also directed at an approach to text classification based on a "bag of phrases" representation; the motivation here being that a phrase contains semantic information that is not present in single keyword. In addition, a given text corpus typically contains many more key-phrase features than keyword features, therefore, providing more potential features to be negated. © Springer-Verlag London Limited 2011.

Cite

CITATION STYLE

APA

Chua, S., Coenen, F., Malcolm, G., Fernando, M., & Constantino, G. (2011). Using negation and phrases in inducing rules for text classification. In Res. and Dev. in Intelligent Syst. XXVIII: Incorporating Applications and Innovations in Intel. Sys. XIX - AI 2011, 31st SGAI Int. Conf. on Innovative Techniques and Applications of Artificial Intel. (pp. 153–166). Springer London. https://doi.org/10.1007/978-1-4471-2318-7_11

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free