Neural discontinuous constituency parsing

15Citations
Citations of this article
83Readers
Mendeley users who have this article in their library.

Abstract

One of the most pressing issues in discontinuous constituency transition-based parsing is that the relevant information for parsing decisions could be located in any part of the stack or the buffer. In this paper, we propose a solution to this problem by replacing the structured perceptron model with a recursive neural model that computes a global representation of the configuration, therefore allowing even the most remote parts of the configuration to influence the parsing decisions. We also provide a detailed analysis of how this representation should be built out of sub-representations of its core elements (words, trees and stack). Additionally, we investigate how different types of swap oracles influence the results. Our model is the first neural discontinuous constituency parser, and it outperforms all the previously published models on three out of four datasets while on the fourth it obtains second place by a tiny difference.

Cite

CITATION STYLE

APA

Stanojević, M., & Alhama, R. G. (2017). Neural discontinuous constituency parsing. In EMNLP 2017 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 1666–1676). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d17-1174

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free