Geometry-driven self-supervised method for 3D human pose estimation

34Citations
Citations of this article
53Readers
Mendeley users who have this article in their library.

Abstract

The neural network based approach for 3D human pose estimation from monocular images has attracted growing interest. However, annotating 3D poses is a labor-intensive and expensive process. In this paper, we propose a novel self-supervised approach to avoid the need of manual annotations. Different from existing weakly/self-supervised methods that require extra unpaired 3D ground-truth data to alleviate the depth ambiguity problem, our method trains the network only relying on geometric knowledge without any additional 3D pose annotations. The proposed method follows the two-stage pipeline: 2D pose estimation and 2D-to-3D pose lifting. We design the transform re-projection loss that is an effective way to explore multi-view consistency for training the 2D-to-3D lifting network. Besides, we adopt the confidences of 2D joints to integrate losses from different views to alleviate the influence of noises caused by the self-occlusion problem. Finally, we design a two-branch training architecture, which helps to preserve the scale information of re-projected 2D poses during training, resulting in accurate 3D pose predictions. We demonstrate the effectiveness of our method on two popular 3D human pose datasets, Human3.6M and MPI-INF-3DHP. The results show that our method significantly outperforms recent weakly/self-supervised approaches.

References Powered by Scopus

Stacked hourglass networks for human pose estimation

3782Citations
N/AReaders
Get full text

Human3.6M: Large scale datasets and predictive methods for 3D human sensing in natural environments

2663Citations
N/AReaders
Get full text

End-to-End Recovery of Human Shape and Pose

1512Citations
N/AReaders
Get full text

Cited by Powered by Scopus

PoseAug: A Differentiable Pose Augmentation Framework for 3D Human Pose Estimation

129Citations
N/AReaders
Get full text

Canonpose: Self-supervised monocular 3D human pose estimation in the wild

89Citations
N/AReaders
Get full text

A comparison review of transfer learning and self-supervised learning: Definitions, applications, advantages and limitations

85Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Li, Y., Li, K., Jiang, S., Zhang, Z., Huang, C., & Da Xu, R. Y. (2020). Geometry-driven self-supervised method for 3D human pose estimation. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 11442–11449). AAAI press. https://doi.org/10.1609/aaai.v34i07.6808

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 26

90%

Professor / Associate Prof. 2

7%

Researcher 1

3%

Readers' Discipline

Tooltip

Computer Science 25

78%

Engineering 5

16%

Chemistry 1

3%

Neuroscience 1

3%

Save time finding and organizing research with Mendeley

Sign up for free