An empirical analysis of optimization for max-margin NLP

13Citations
Citations of this article
108Readers
Mendeley users who have this article in their library.

Abstract

Despite the convexity of structured maxmargin objectives (Taskar et al., 2004; Tsochantaridis et al., 2004), the many ways to optimize them are not equally effective in practice. We compare a range of online optimization methods over a variety of structured NLP tasks (coreference, summarization, parsing, etc) and find several broad trends. First, margin methods do tend to outperform both likelihood and the perceptron. Second, for max-margin objectives, primal optimization methods are often more robust and progress faster than dual methods. This advantage is most pronounced for tasks with dense or continuous-valued features. Overall, we argue for a particularly simple online primal subgradient descent method that, despite being rarely mentioned in the literature, is surprisingly effective in relation to its alternatives.

Cite

CITATION STYLE

APA

Kummerfeld, J. K., Berg-Kirkpatrick, T., & Klein, D. (2015). An empirical analysis of optimization for max-margin NLP. In Conference Proceedings - EMNLP 2015: Conference on Empirical Methods in Natural Language Processing (pp. 273–279). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d15-1032

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free