Abstract
Using the Internet for the collection of data is quite common these days. This process is called crowdsourcing and enables the collection of large amounts of data at reasonable costs. While being an inexpensive method, this data typically is of lower quality. Filtering data sets is therefore required. The occurring errors can be classified into different groups. There are technical issues and human errors. For speech recording, technical issues could be a noisy background. Human errors arise when the task is misunderstood. We employ several techniques for recognizing errors and eliminating faulty data sets in user input data for a Spoken Dialog System (SDS). Furthermore, we compare three different kinds of questionnaires (QNRs) for a given set of seven tasks. We analyze the characteristics of the resulting data sets and give a recommendation which type of QNR might be the most suitable one for a given purpose.
Cite
CITATION STYLE
Schmidt, M., Müller, M., Wagner, M., Stüker, S., Waibel, A., Hofmann, H., & Werner, S. (2015). Evaluation of crowdsourced user input data for spoken dialog systems. In SIGDIAL 2015 - 16th Annual Meeting of the Special Interest Group on Discourse and Dialogue, Proceedings of the Conference (pp. 427–431). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w15-4657
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.