On the class distribution labelling step sensitivity of CO-TRAINING

1Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

CO-TRAINING can learn from datasets having a small number of labelled examples and a large number of unlabelled ones. It is an iterative algorithm where examples labelled in previous iterations are used to improve the classification of examples from the unlabelled set. However, as the number of initial labelled examples is often small we do not have reliable estimates regarding the underlying population which generated the data. In this work we make the claim that the proportion in which examples are labelled is a key parameter to CO-TRAINING. Furthermore, we have done a series of experiments to investigate how the proportion in which we label examples in each step influences CO-TRAINING performance. Results show that CO-TRAINING should be used with care in challenging domains. © 2006 International Federation for Information Processing.

Cite

CITATION STYLE

APA

Matsubara, E. T., Monard, M. C., & Prati, R. C. (2006). On the class distribution labelling step sensitivity of CO-TRAINING. IFIP International Federation for Information Processing, 217, 199–208. https://doi.org/10.1007/978-0-387-34747-9_21

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free