Abstract
We present a discriminative, latent variable approach to syntactic parsing in which rules exist at multiple scales of refinement. The model is formally a latent variable CRF grammar over trees, learned by iteratively splitting grammar productions (not categories). Different regions of the grammar are refined to different degrees, yielding grammars which are three orders of magnitude smaller than the single-scale baseline and 20 times smaller than the split-and-merge grammars of Petrov et al. (2006). In addition, our discriminative approach integrally admits features beyond local tree configurations. We present a multi-scale training method along with an efficient CKY-style dynamic program. On a variety of domains and languages, this method produces the best published parsing accuracies with the smallest reported grammars. © 2008 Association for Computational Linguistics.
Cite
CITATION STYLE
Petrov, S., & Klein, D. (2008). Sparse multi-scale grammars for discriminative latent variable parsing. In EMNLP 2008 - 2008 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference: A Meeting of SIGDAT, a Special Interest Group of the ACL (pp. 867–876). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1613715.1613827
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.