In this paper we investigate whether a combination of statistical, neural network and cache language models can outperform a basic statistical model. These models have been developed, tested and exploited for a Czech spontaneous speech data, which is very different from common written Czech and is specified by a small set of the data available and high inflection of the words. As a baseline model we used a trigram model and after its training several cache models interpolated with the baseline model have been tested and measured on a perplexity. Finally, an evaluation of the model with the lowest perplexity has been performed on speech recordings of phone calls. © 2012 Springer-Verlag.
CITATION STYLE
Soutner, D., Loose, Z., Müller, L., & Pražák, A. (2012). Neural network language model with cache. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7499 LNAI, pp. 528–534). https://doi.org/10.1007/978-3-642-32790-2_64
Mendeley helps you to discover research relevant for your work.