Partial training for a lexicalized-grammar parser

14Citations
Citations of this article
85Readers
Mendeley users who have this article in their library.

Abstract

We propose a solution to the annotation bottleneck for statistical parsing, by exploiting the lexicalized nature of Combinatory Categorial Grammar (CCG). The parsing model uses predicate-argument dependencies for training, which are derived from sequences of CCG lexical categories rather than full derivations. A simple method is used for extracting dependencies from lexical category sequences, resulting in high precision, yet incomplete and noisy data. The dependency parsing model of Clark and Curran (2004b) is extended to exploit this partial training data. Remarkably, the accuracy of the parser trained on data derived from category sequences alone is only 1.3% worse in terms of F-score than the parser trained on complete dependency structures. © 2006 Association for Computational Linguistics.

Cite

CITATION STYLE

APA

Clark, S., & Curran, J. R. (2006). Partial training for a lexicalized-grammar parser. In HLT-NAACL 2006 - Human Language Technology Conference of the North American Chapter of the Association of Computational Linguistics, Proceedings of the Main Conference (pp. 144–151). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1220835.1220854

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free