CO-TRAINING can learn from datasets having a small number of labelled examples and a large number of unlabelled ones. It is an iterative algorithm where examples labelled in previous iterations are used to improve the classification of examples from the unlabelled set. However, as the number of initial labelled examples is often small we do not have reliable estimates regarding the underlying population which generated the data. In this work we make the claim that the proportion in which examples are labelled is a key parameter to CO-TRAINING. Furthermore, we have done a series of experiments to investigate how the proportion in which we label examples in each step influences CO-TRAINING performance. Results show that CO-TRAINING should be used with care in challenging domains. © 2006 International Federation for Information Processing.
CITATION STYLE
Matsubara, E. T., Monard, M. C., & Prati, R. C. (2006). On the class distribution labelling step sensitivity of CO-TRAINING. IFIP International Federation for Information Processing, 217, 199–208. https://doi.org/10.1007/978-0-387-34747-9_21
Mendeley helps you to discover research relevant for your work.