Improving active learning by avoiding ambiguous samples

4Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

If label information in a classification task is expensive, it can be beneficial to use active learning to get the most informative samples to label by a human. However, there can be samples which are meaningless to the human or recorded wrongly. If these samples are near the classifier’s decision boundary, they are queried repeatedly for labeling. This is inefficient for training because the human can not label these samples correctly and this may lower human acceptance. We introduce an approach to compensate the problem of ambiguous samples by excluding clustered samples from labeling. We compare this approach to other state-of-the-art methods. We further show that we can improve the accuracy in active learning and reduce the number of ambiguous samples queried while training.

Cite

CITATION STYLE

APA

Limberg, C., Wersing, H., & Ritter, H. (2018). Improving active learning by avoiding ambiguous samples. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11139 LNCS, pp. 518–527). Springer Verlag. https://doi.org/10.1007/978-3-030-01418-6_51

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free