Fine-Grained Egocentric Hand-Object Segmentation: Dataset, Model, and Applications

13Citations
Citations of this article
27Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Egocentric videos offer fine-grained information for high-fidelity modeling of human behaviors. Hands and interacting objects are one crucial aspect of understanding a viewer’s behaviors and intentions. We provide a labeled dataset consisting of 11,243 egocentric images with per-pixel segmentation labels of hands and objects being interacted with during a diverse array of daily activities. Our dataset is the first to label detailed hand-object contact boundaries. We introduce a context-aware compositional data augmentation technique to adapt to out-of-distribution YouTube egocentric video. We show that our robust hand-object segmentation model and dataset can serve as a foundational tool to boost or enable several downstream vision applications, including hand state classification, video activity recognition, 3D mesh reconstruction of hand-object interactions, and video inpainting of hand-object foregrounds in egocentric videos. Dataset and code are available at: https://github.com/owenzlz/EgoHOS.

Cite

CITATION STYLE

APA

Zhang, L., Zhou, S., Stent, S., & Shi, J. (2022). Fine-Grained Egocentric Hand-Object Segmentation: Dataset, Model, and Applications. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 13689 LNCS, pp. 127–145). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-031-19818-2_8

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free