Abstract
This work addresses the task of dependency labeling - assigning labels to an (unlabeled) dependency tree. We employ and extend a feature representation learning approach, optimizing it for both high speed and accuracy. We apply our labeling model on top of state-of-the-art parsers and evaluate its performance on standard benchmarks including the CoNLL-2009 and the English PTB datasets. Our model processes over 1,700 English sentences per second, which is 30 times faster than the sparse-feature method. It improves labeling accuracy over the outputs of top parsers, achieving the best LAS on 5 out of 7 datasets1.
Cite
CITATION STYLE
Shen, T., Lei, T., & Barzilay, R. (2016). Making dependency labeling simple, fast and accurate. In 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL HLT 2016 - Proceedings of the Conference (pp. 1089–1094). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/n16-1126
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.