Patchit: Self-supervised network weight initialization for fine-grained recognition

10Citations
Citations of this article
23Readers
Mendeley users who have this article in their library.
Get full text

Abstract

ConvNet training is highly sensitive to initialization of the weights. A widespread approach is to initialize the network with weights trained for a different task, an auxiliary task. The ImageNet-based ILSVRC classification task is a very popular choice for this, as it has shown to produce powerful feature representations applicable to a wide variety of tasks. However, this creates a significant entry barrier to exploring non-standard architectures. In this paper, we propose a self-supervised pretraining, the PatchTask, to obtain weight initializations for fine-grained recognition problems, such as person attribute recognition, pose estimation, or action recognition. Our pretraining allows us to leverage additional unlabeled data from the same source, which is often readily available, such as detection bounding boxes. We experimentally show that our method outperforms a standard random initialization by a considerable margin and closely matches the ImageNet-based initialization.

Cite

CITATION STYLE

APA

Sudowe, P., & Leibe, B. (2016). Patchit: Self-supervised network weight initialization for fine-grained recognition. In British Machine Vision Conference 2016, BMVC 2016 (Vol. 2016-September, pp. 2266–2270). British Machine Vision Conference, BMVC. https://doi.org/10.5244/C.30.75

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free