A recurrent neural model with attention for the recognition of Chinese implicit discourse relations

37Citations
Citations of this article
104Readers
Mendeley users who have this article in their library.

Abstract

We introduce an attention-based Bi-LSTM for Chinese implicit discourse relations and demonstrate that modeling argument pairs as a joint sequence can outperform word order-agnostic approaches. Our model benefits from a partial sampling scheme and is conceptually simple, yet achieves state-of-the-art performance on the Chinese Discourse Treebank. We also visualize its attention activity to illustrate the model’s ability to selectively focus on the relevant parts of an input sequence.

Cite

CITATION STYLE

APA

Rönnqvist, S., Schenk, N., & Chiarcos, C. (2017). A recurrent neural model with attention for the recognition of Chinese implicit discourse relations. In ACL 2017 - 55th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 2, pp. 256–262). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/P17-2040

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free