MobileTutAR: a Lightweight Augmented Reality Tutorial System using Spatially Situated Human Segmentation Videos

8Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Augmented reality (AR) tutorial systems have a strong potential to help workers improve learning efficiency in the ongoing trend of Industry 4.0. The current state-of-the-art point cloud approach usually requires cumbersome preparation and additional computational resources only to suffer from low-resolution visual results. To overcome the above limitations, we propose MobileTutAR, a lightweight AR tutorial system that runs entirely on mobile devices providing high definition spatially situated tutorial videos. Our approach captures the tutorial images that extract a human's body segmentation combined with a user-defined area of interest. The system then projects the tutoring content spatially in situ, to align with the human expert's recorded position. When playing back, the learner is guided by a navigation-centered user interface to observe the segmentation video from the recorder's original position/orientation. In this way, we deliver a high-definition AR experience without needing any cumbersome equipment, exotic computational resources, or in-depth training.

Cite

CITATION STYLE

APA

Cao, Y., Fuste, A., & Heun, V. (2022). MobileTutAR: a Lightweight Augmented Reality Tutorial System using Spatially Situated Human Segmentation Videos. In Conference on Human Factors in Computing Systems - Proceedings. Association for Computing Machinery. https://doi.org/10.1145/3491101.3519639

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free