Crowdsourcing offers a convenient means of obtaining labeled data quickly and inexpensively. However, crowdsourced labels are often noisier than expert-annotated data, making it difficult to aggregate them meaningfully. We present an aggregation approach that learns a regression model from crowdsourced annotations to predict aggregated labels for instances that have no expert adjudications. The predicted labels achieve a correlation of 0.594 with expert labels on our data, outperforming the best alternative aggregation method by 11.9%. Our approach also outperforms the alternatives on third-party datasets.
CITATION STYLE
Parde, N., & Nielsen, R. D. (2017). Finding patterns in noisy crowds: Regression-based annotation aggregation for crowdsourced data. In EMNLP 2017 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 1907–1912). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d17-1204
Mendeley helps you to discover research relevant for your work.