Performance of a SCFG-based language model with training data sets of increasing size

1Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper, a hybrid language model which combines a word-based n-gram and a category-based Stochastic Context-Free Grammar (SCFG) is evaluated for training data sets of increasing size. Different estimation algorithms for learning SCFGs in General Format and in Chomsky Normal Form are considered. Experiments on the UPenn Treebank corpus are reported. These experiments have been carried out in terms of the test set perplexity and the word error rate in a speech recognition experiment. © Springer-Verlag Berlin Heidelberg 2005.

Cite

CITATION STYLE

APA

Sánchez, J. A., Benedí, J. M., & Linares, D. (2005). Performance of a SCFG-based language model with training data sets of increasing size. In Lecture Notes in Computer Science (Vol. 3523, pp. 586–594). Springer Verlag. https://doi.org/10.1007/11492542_72

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free