Abstract
This paper demonstrates how unsupervised techniques can be used to learn models of deep linguistic structure. Determining the semantic roles of a verb's dependents is an important step in natural language understanding. We present a method for learning models of verb argument patterns directly from unannotated text. The learned models are similar to existing verb lexicons such as VerbNet and PropBank, but additionally include statistics about the linkings used by each verb. The method is based on a structured probabilistic model of the domain, and unsupervised learning is performed with the EM algorithm. The learned models can also be used discriminatively as semantic role labelers, and when evaluated relative to the PropBank annotation, the best learned model reduces 28% of the error between an informed baseline and an oracle upper bound. © 2006 Association for Computational Linguistics.
Cite
CITATION STYLE
Grenager, T., & Manning, C. D. (2006). Unsupervised discovery of a statistical verb lexicon. In COLING/ACL 2006 - EMNLP 2006: 2006 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference (pp. 1–8). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1610075.1610077
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.