Accelerated training of conditional random fields with stochastic gradient methods

81Citations
Citations of this article
215Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We apply Stochastic Meta-Descent (SMD), a stochastic gradient optimization method with gain vector adaptation, to the training of Conditional Random Fields (CRFs). On several large data sets, the resulting optimizer converges to the same quality of solution over an order of magnitude faster than limited-memory BFGS, the leading method reported to date. We report results for both exact and inexact inference techniques.

Cite

CITATION STYLE

APA

Vishwanathan, S. V. N., Schraudolph, N. N., Schmidt, M. W., & Murphy, K. P. (2006). Accelerated training of conditional random fields with stochastic gradient methods. In ACM International Conference Proceeding Series (Vol. 148, pp. 969–976). https://doi.org/10.1145/1143844.1143966

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free