Robust text classification using a hysteresis-driven extended SRN

1Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Recurrent Neural Network (RNN) models have been shown to perform well on artificial grammars for sequential classification tasks over long-term time-dependencies. However, there is a distinct lack of the application of RNNs to real-world text classification tasks. This paper presents results on the capabilities of extended two-context layer SRN models (xRNN) applied to the classification of the Reuters-21578 corpus. The results show that the introduction of high levels of noise to sequences of words in titles, where noise is defined as the unimportant stopwords found in natural language text, is very robustly handled by the classifiers which maintain consistent levels of performance. Comparisons are made with SRN and MLP models, as well as other existing classifiers for the text classification task. © Springer-Verlag Berlin Heidelberg 2007.

Cite

CITATION STYLE

APA

Arevian, G., & Panchev, C. (2007). Robust text classification using a hysteresis-driven extended SRN. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 4669 LNCS, pp. 425–434). Springer Verlag. https://doi.org/10.1007/978-3-540-74695-9_44

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free