Annealing techniques for unsupervised statistical language learning

28Citations
Citations of this article
95Readers
Mendeley users who have this article in their library.

Abstract

Exploiting unannotated natural language data is hard largely because unsupervised parameter estimation is hard. We describe deterministic annealing (Rose et al., 1990) as an appealing alternative to the Expectation-Maximization algorithm (Dempster et al., 1977). Seeking to avoid search error, DA begins by globally maximizing an easy concave function and maintains a local maximum as it gradually morphs the function into the desired non-concave likelihood function. Applying DA to parsing and tagging models is shown to be straightforward; significant improvements over EM are shown on a part-of-speech tagging task. We describe a variant, skewed DA, which can incorporate a good initializer when it is available, and show significant improvements over EM on a grammar induction task.

References Powered by Scopus

Elements of Information Theory

36599Citations
N/AReaders
Get full text

Optimization by simulated annealing

34738Citations
N/AReaders
Get full text

Deterministic annealing for clustering, compression, classification, regression, and related optimization problems

649Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Minimum risk annealing for training log-linear models

102Citations
N/AReaders
Get full text

Linguistic Structure Prediction

59Citations
N/AReaders
Get full text

On the link between gaussian homotopy continuation and convex envelopes

48Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Smith, N. A., & Eisner, J. (2004). Annealing techniques for unsupervised statistical language learning. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 486–493). Association for Computational Linguistics (ACL). https://doi.org/10.3115/1218955.1219017

Readers over time

‘09‘10‘11‘12‘13‘14‘15‘16‘17‘18‘19‘20‘21‘22‘23‘24‘2507142128

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 31

58%

Professor / Associate Prof. 10

19%

Researcher 9

17%

Lecturer / Post doc 3

6%

Readers' Discipline

Tooltip

Computer Science 37

69%

Linguistics 10

19%

Engineering 5

9%

Social Sciences 2

4%

Save time finding and organizing research with Mendeley

Sign up for free
0