Learning connective-based word representations for implicit discourse relation identification

32Citations
Citations of this article
104Readers
Mendeley users who have this article in their library.

Abstract

We introduce a simple semi-supervised approach to improve implicit discourse relation identification. This approach harnesses large amounts of automatically extracted discourse connectives along with their arguments to construct new distributional word representations. Specifically, we represent words in the space of discourse connectives as a way to directly encode their rhetorical function. Experiments on the Penn Discourse Treebank demonstrate the effectiveness of these task-tailored representations in predicting implicit discourse relations. Our results indeed show that, despite their simplicity, these connective-based representations outperform various off-the-shelf word embeddings, and achieve state-of-the-art performance on this problem.

Cite

CITATION STYLE

APA

Braud, C., & Denis, P. (2016). Learning connective-based word representations for implicit discourse relation identification. In EMNLP 2016 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 203–213). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d16-1020

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free