Abstract
We outline a learning framework that aims at identifying useful contextual cues for knowledge-based word sense disambiguation. The usefulness of individual context words is evaluated based on diverse lexico-statistical and syntactic information, as well as simple word distance. Experiments using two different knowledge-based methods and benchmark datasets show significant improvements due to context modeling, beating the conventional window-based approach.
Cite
CITATION STYLE
Wasserman-Pritsker, E., Cohen, W. W., & Minkov, E. (2015). Learning to identify the best contexts for knowledge-based WSD. In Conference Proceedings - EMNLP 2015: Conference on Empirical Methods in Natural Language Processing (pp. 1662–1667). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d15-1192
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.