Abstract
Algorithms which are based on modeling speech as a finite-state, hidden Markov process have been very successful in recent years. This paper presents a generalization of these algorithms to certain denumerable-state, hidden Markov processes. This algorithm permits automatic training of the stochastic analog of an arbitrary context free grammar. In particular, in contrast to many grammatical inference methods, the new algorithm allows the grammar to have an arbitrary degree of ambiguity. Since natural language is often syntactically ambiguous, it is necessary for the grammatical inference algorithm to allow for this ambiguity. Furthermore, allowing ambiguity in the grammar allows errors in the recognition process to be explicitly modeled in the grammar rather than added as an extra component.
Cite
CITATION STYLE
Baker, J. K. (1979). Trainable grammars for speech recognition. The Journal of the Acoustical Society of America, 65(S1), S132–S132. https://doi.org/10.1121/1.2017061
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.