Learning word representations with cross-sentence dependency for end-to-end Co-reference resolution

10Citations
Citations of this article
99Readers
Mendeley users who have this article in their library.

Abstract

In this work, we present a word embedding model that learns cross-sentence dependency for improving end-to-end co-reference resolution (E2E-CR). While the traditional E2ECR model generates word representations by running long short-term memory (LSTM) recurrent neural networks on each sentence of an input article or conversation separately, we propose linear sentence linking and attentional sentence linking models to learn cross-sentence dependency. Both sentence linking strategies enable the LSTMs to make use of valuable information from context sentences while calculating the representation of the current input word. With this approach, the LSTMs learn word embeddings considering knowledge not only from the current sentence but also from the entire input document. Experiments show that learning cross-sentence dependency enriches information contained by the word representations, and improves the performance of the co-reference resolution model compared with our baseline.

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Luo, H., & Glass, J. (2018). Learning word representations with cross-sentence dependency for end-to-end Co-reference resolution. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018 (pp. 4829–4833). Association for Computational Linguistics. https://doi.org/10.18653/v1/d18-1518

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 30

71%

Researcher 8

19%

Professor / Associate Prof. 2

5%

Lecturer / Post doc 2

5%

Readers' Discipline

Tooltip

Computer Science 38

76%

Linguistics 5

10%

Engineering 4

8%

Neuroscience 3

6%

Save time finding and organizing research with Mendeley

Sign up for free