Convolutions Are All You Need (For Classifying Character Sequences)

10Citations
Citations of this article
79Readers
Mendeley users who have this article in their library.

Abstract

While recurrent neural networks (RNNs) are widely used for text classification, they demonstrate poor performance and slow convergence when trained on long sequences. When text is modeled as characters instead of words, the longer sequences make RNNs a poor choice. Convolutional neural networks (CNNs), although somewhat less ubiquitous than RNNs, have an internal structure more appropriate for long-distance character dependencies. To better understand how CNNs and RNNs differ in handling long sequences, we use them for text classification tasks in several character-level social media datasets. The CNN models vastly outperform the RNN models in our experiments, suggesting that CNNs are superior to RNNs at learning to classify character-level data.

Cite

CITATION STYLE

APA

Wood-Doughty, Z., Andrews, N., & Dredze, M. (2018). Convolutions Are All You Need (For Classifying Character Sequences). In 4th Workshop on Noisy User-Generated Text, W-NUT 2018 - Proceedings of the Workshop (pp. 208–213). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w18-6127

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free