Towards Digital Twins of 3D Reconstructed Apparel Models with an End-to-End Mobile Visualization

11Citations
Citations of this article
30Readers
Mendeley users who have this article in their library.

Abstract

Digital twin technologies are still developing and are being increasingly leveraged to facilitate daily life activities. This study presents a novel approach for leveraging the capability of mobile devices for photo collection, cloud processing, and deep learning-based 3D generation, with seamless display in virtual reality (VR) wearables. The purpose of our study is to provide a system that makes use of cloud computing resources to offload the resource-intensive activities of 3D reconstruction and deep-learning-based scene interpretation. We establish an end-to-end pipeline from 2D to 3D reconstruction, which automatically builds accurate 3D models from collected photographs using sophisticated deep-learning techniques. These models are then converted to a VR-compatible format, allowing for immersive and interactive experiences on wearable devices. Our findings attest to the completion of 3D entities regenerated by the CAP–UDF model using ShapeNetCars and Deep Fashion 3D datasets with a discrepancy in L2 Chamfer distance of only 0.089 and 0.129, respectively. Furthermore, the demonstration of the end-to-end process from 2D capture to 3D visualization on VR occurs continuously.

Cite

CITATION STYLE

APA

Doungtap, S., Petchhan, J., Phanichraksaphong, V., & Wang, J. H. (2023). Towards Digital Twins of 3D Reconstructed Apparel Models with an End-to-End Mobile Visualization. Applied Sciences (Switzerland), 13(15). https://doi.org/10.3390/app13158571

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free