LIPREADING USING SHAPE, SHADING AND SCALE

24Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper compares three methods of lipreading for visual and audio-visual speech recognition. Lip shape information is obtained using an Active Shape Model (ASM) lip tracker but is not as effective as modelling the combined shape and enclosed greylevel surface using an Active Appearance Model (AAM). A non-tracked alternative is a nonlinear transform of the image using a multiscale spatial analysis (MSA). This performs almost identically to AAM’s in both visual and audio-visual recognition tasks on a multi-talker database of isolated letters.

Cite

CITATION STYLE

APA

Matthews, I., Cootes, T., Cox, S., Harvey, R., & Bangham, J. A. (1998). LIPREADING USING SHAPE, SHADING AND SCALE. In Auditory-Visual Speech Processing 1998, AVSP 1998 (pp. 73–78). The International Society for Computers and Their Applications (ISCA). https://doi.org/10.25144/18965

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free