Abstract
Distributional semantic models can predict many linguistic phenomena, including word similarity, lexical ambiguity, and semantic priming, or even to pass TOEFL synonymy and analogy tests (Landauer and Dumais, 1997; Griffiths et al., 2007; Turney and Pantel, 2010). But what does it take to create a competitive distributional model? Levy et al. (2015) argue that the key to success lies in hyperparameter tuning rather than in the model's architecture. More hyperparameters trivially lead to potential performance gains, but what do they actually do to improve the models? Are individual hyperparameters' contributions independent of each other? Or are only specific parameter combinations beneficial? To answer these questions, we perform a quantitative and qualitative evaluation of major hyperparameters as identified in previous research.
Cite
CITATION STYLE
Krebs, A., & Paperno, D. (2016). When hyperparameters help: Beneficial parameter combinations in distributional semantic models. In *SEM 2016 - 5th Joint Conference on Lexical and Computational Semantics, Proceedings (pp. 97–101). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/s16-2011
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.