With the growing of crowdsourcing services, gathering training data for supervised machine learning has become cheaper and faster than engaging experts. However, the quality of the crowd-generated labels remains an open issue. This is basically due to the wide ranging expertise levels of the participants in the labeling process. In this paper, we present an iterative approach of label aggregation based on the belief function theory that simultanously estimates labels, the reliability of participants and difficulty of each task. Our empirical evaluation demonstrate the efficiency of our method as it gives better quality labels.
CITATION STYLE
Abassi, L., & Boukhris, I. (2017). Iterative aggregation of crowdsourced tasks within the belief function theory. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 10369 LNAI, pp. 159–168). Springer Verlag. https://doi.org/10.1007/978-3-319-61581-3_15
Mendeley helps you to discover research relevant for your work.