Explainable modeling of annotations in crowdsourcing

7Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Aggregation models for improving the quality of annotations collected via crowdsourcing have been widely studied, but far less has been done to explain why annotators make the mistakes that they do. To this end, we propose a joint aggregation and worker clustering model that detects patterns underlying crowd worker labels to characterize varieties of labeling errors. We evaluate our approach on a Named Entity Recognition dataset labeled by Mechanical Turk workers in both a retrospective experiment and a small human study. The former shows that our joint model improves the quality of clusters vs. aggregation followed by clustering. Results of the latter suggest that clusters aid human sense-making in interpreting worker labels and predicting worker mistakes. By enabling better explanation of annotator mistakes, our model creates a new opportunity to help Requesters improve task instructions and to help crowd annotators learn from their mistakes. Source code, data, and supplementary material is shared online.

Author supplied keywords

Cite

CITATION STYLE

APA

Nguyen, A. T., Lease, M., & Wallace, B. C. (2019). Explainable modeling of annotations in crowdsourcing. In International Conference on Intelligent User Interfaces, Proceedings IUI (Vol. Part F147615, pp. 575–579). Association for Computing Machinery. https://doi.org/10.1145/3301275.3302276

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free