Curriculum Labeling: Revisiting Pseudo-Labeling for Semi-Supervised Learning

206Citations
Citations of this article
265Readers
Mendeley users who have this article in their library.

Abstract

In this paper we revisit the idea of pseudo-labeling in the context of semi-supervised learning where a learning algorithm has access to a small set of labeled samples and a large set of unlabeled samples. Pseudo-labeling works by applying pseudo-labels to samples in the unlabeled set by using a model trained on the combination of the labeled samples and any previously pseudo-labeled samples, and iteratively repeating this process in a self-training cycle. Current methods seem to have abandoned this approach in favor of consistency regularization methods that train models under a combination of different styles of self-supervised losses on the unlabeled samples and standard supervised losses on the labeled samples. We empirically demonstrate that pseudo-labeling can in fact be competitive with the state-of-the-art, while being more resilient to out-of-distribution samples in the unlabeled set. We identify two key factors that allow pseudo-labeling to achieve such remarkable results (1) applying curriculum learning principles and (2) avoiding concept drift by restarting model parameters before each self-training cycle. We obtain 94.91% accuracy on CIFAR-10 using only 4, 000 labeled samples, and 68.87% top-1 accuracy on Imagenet-ILSVRC using only 10% of the labeled samples.

References Powered by Scopus

ImageNet: A Large-Scale Hierarchical Image Database

50793Citations
N/AReaders
Get full text

ImageNet Large Scale Visual Recognition Challenge

30411Citations
N/AReaders
Get full text

A survey on Image Data Augmentation for Deep Learning

8152Citations
N/AReaders
Get full text

Cited by Powered by Scopus

ST++: Make Self-trainingWork Better for Semi-supervised Semantic Segmentation

285Citations
N/AReaders
Get full text

DMT: Dynamic mutual training for semi-supervised learning

119Citations
N/AReaders
Get full text

MCF: Mutual Correction Framework for Semi-Supervised Medical Image Segmentation

75Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Cascante-Bonilla, P., Tan, F., Qi, Y., & Ordonez, V. (2021). Curriculum Labeling: Revisiting Pseudo-Labeling for Semi-Supervised Learning. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 8A, pp. 6912–6920). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i8.16852

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 112

73%

Researcher 29

19%

Lecturer / Post doc 7

5%

Professor / Associate Prof. 6

4%

Readers' Discipline

Tooltip

Computer Science 122

79%

Engineering 27

17%

Environmental Science 3

2%

Economics, Econometrics and Finance 3

2%

Save time finding and organizing research with Mendeley

Sign up for free