Wav2KWS: Transfer Learning from Speech Representations for Keyword Spotting

41Citations
Citations of this article
47Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

With the expanding development of on-device artificial intelligence, voice-enabled devices such as smart speakers, wearables, and other on-device or edge processing systems have been proposed. However, building or obtaining large training datasets that are essential for robust keyword spotting (KWS) remains cumbersome. To address this problem, we propose a deep neural network that can rapidly establish a high-performance KWS system from arbitrary keyword instruction sets. We use an encoder pretrained with a large-scale speech corpus as the backbone network and then design an effective transfer network for KWS. To demonstrate the feasibility of the proposed network, various experiments were conducted on Google Speech Command Datasets V1 and V2. In addition, to verify the applicability of the network for different languages, we conducted experiments using three different Korean speech command datasets. The proposed network outperforms state-of-the-art deep neural networks in both experiments. Furthermore, the proposed network can understand real human voice even when trained with synthetic text-to-speech data.

Cite

CITATION STYLE

APA

Seo, D., Oh, H. S., & Jung, Y. (2021). Wav2KWS: Transfer Learning from Speech Representations for Keyword Spotting. IEEE Access, 9, 80682–80691. https://doi.org/10.1109/ACCESS.2021.3078715

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free