Unsupervised labeled parsing with deep inside-outside recursive autoencoders

10Citations
Citations of this article
77Readers
Mendeley users who have this article in their library.

Abstract

Understanding text often requires identifying meaningful constituent spans such as noun phrases and verb phrases. In this work, we show that we can effectively recover these types of labels using the learned phrase vectors from deep inside-outside recursive autoencoders (DIORA). Specifically, we cluster span representations to induce span labels. Additionally, we improve the model's labeling accuracy by integrating latent code learning into the training procedure. We evaluate this approach empirically through unsupervised labeled constituency parsing. Our method outperforms ELMo and BERT on two versions of the Wall Street Journal (WSJ) dataset and is competitive to prior work that requires additional human annotations, improving over a previous state-of-the-art system that depends on ground-truth part-of-speech tags by 5 absolute F1 points (19% relative error reduction).

Cite

CITATION STYLE

APA

Drozdov, A., Verga, P., Chen, Y. P., Iyyer, M., & McCallum, A. (2019). Unsupervised labeled parsing with deep inside-outside recursive autoencoders. In EMNLP-IJCNLP 2019 - 2019 Conference on Empirical Methods in Natural Language Processing and 9th International Joint Conference on Natural Language Processing, Proceedings of the Conference (pp. 1507–1512). Association for Computational Linguistics. https://doi.org/10.18653/v1/d19-1161

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free