Preferred appearance of captions generated by automatic speech recognition for deaf and hard-of-hearing viewers

39Citations
Citations of this article
34Readers
Mendeley users who have this article in their library.
Get full text

Abstract

As the accuracy of Automatic Speech Recognition (ASR) nears human-level quality, it might become feasible as an accessibility tool for people who are Deaf and Hard of Hearing (DHH) to transcribe spoken language to text. We conducted a study using in-person laboratory methodologies, to investigate requirements and preferences for new ASR-based captioning services when used in a small group meeting context. The open-ended comments reveal an interesting dynamic between: caption readability (visibility of text) and occlusion (captions blocking the video contents). Our 105 DHH participants provided valuable feedback on a variety of caption-appearance parameters (strongly preferring familiar styles such as closed captions), and in this paper we start a discussion on how ASR captioning could be visually styled to improve text readability for DHH viewers.

Cite

CITATION STYLE

APA

Berke, L., Seita, M., Albusays, K., & Huenerfauth, M. (2019). Preferred appearance of captions generated by automatic speech recognition for deaf and hard-of-hearing viewers. In Conference on Human Factors in Computing Systems - Proceedings. Association for Computing Machinery. https://doi.org/10.1145/3290607.3312921

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free