Sign up & Download
Sign in

Neural network models of sensory integration for improved vowel recognition

by Ben P. Yuhas, Moise H. Goldstein, Terrence J. Sejnowski, Robert E. Jenkins
Proceedings of the IEEE ()

Abstract

It is demonstrated that multiple sources of speech information can be integrated at a subsymbolic level to improve vowel recognition. Feedforward and recurrent neural networks are trained to estimate the acoustic characteristics of a vocal tract from images of the speaker's mouth. These estimates are then combined with the noise-degraded acoustic information, effectively increasing the signal-to-noise ratio and improving the recognition of these noise-degraded signals. Alternative symbolic strategies such as direct categorization of the visual signals into vowels are also presented. The performances of these neural networks compare favorably with human performance and with other pattern-matching and estimation techniques

Cite this document (BETA)

Readership Statistics

16 Readers on Mendeley
by Discipline
 
 
 
by Academic Status
 
25% Ph.D. Student
 
13% Assistant Professor
 
13% Professor
by Country
 
6% Sweden
 
6% United Kingdom
 
6% Poland

Sign up today - FREE

Mendeley saves you time finding and organizing research. Learn more

  • All your research in one place
  • Add and import papers easily
  • Access it anywhere, anytime

Start using Mendeley in seconds!

Already have an account? Sign in