In this paper, we propose to learn a powerful Re-ID model by using less labeled data together with lots of unlabeled data, i.e. semi-supervised Re-ID. Such kind of learning enables Re-ID model to be more generalizable and scalable to real-world scenes. Specifically, we design a two-stream encoder-decoder-based structure with shared modules and parameters. For the encoder module, we take the original person image with its horizontal mirror image as a pair of inputs and encode deep features with identity and structural information properly disentangled. Then different combinations of disentangling features are used to reconstruct images in the decoder module. In addition to the commonly used constraints from identity consistency and image reconstruction consistency for loss function definition, we design a novel loss function of enforcing consistent transformation constraints on disentangled features. It is free of labels, and can be applied to both supervised and unsupervised learning branches in our model. Extensive results on four Re-ID datasets demonstrate that by reducing 5/6 labeled data, Our method achieves the best performance on Market-1501 and CUHK03, and comparable accuracy on DukeMTMC-reID and MSMT17.
CITATION STYLE
Hao, G., Yang, Y., Zhou, X., Wang, G., & Lei, Z. (2021). Horizontal Flipping Assisted Disentangled Feature Learning for Semi-supervised Person Re-identification. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12624 LNCS, pp. 21–37). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-69535-4_2
Mendeley helps you to discover research relevant for your work.