Hierarchical Convolutional Attention Networks for Text Classification

56Citations
Citations of this article
151Readers
Mendeley users who have this article in their library.

Abstract

Recent work in machine translation has demonstrated that self-attention mechanisms can be used in place of recurrent neural networks to increase training speed without sacrificing model accuracy. We propose combining this approach with the benefits of convolutional filters and a hierarchical structure to create a document classification model that is both highly accurate and fast to train - we name our method Hierarchical Convolutional Attention Networks. We demonstrate the effectiveness of this architecture by surpassing the accuracy of the current state-of-the-art on several classification tasks while being twice as fast to train.

Cite

CITATION STYLE

APA

Gao, S., Ramanathan, A., & Tourassi, G. (2018). Hierarchical Convolutional Attention Networks for Text Classification. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 11–23). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w18-3002

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free