Abstract
3D models derived from point clouds are useful in various shapes to optimize the trade-off between precision and geometric complexity. They are defined at different granularity levels according to each indoor situation. In this article, we present an integrated 3D semantic reconstruction framework that leverages segmented point cloud data and domain ontologies. Our approach follows a part-to-whole conception which models a point cloud in parametric elements usable per instance and aggregated to obtain a global 3D model. We first extract analytic features, object relationships and contextual information to permit better object characterization. Then, we propose a multi-representation modelling mechanism augmented by automatic recognition and fitting from the 3D library ModelNet10 to provide the best candidates for several 3D scans of furniture. Finally, we combine every element to obtain a consistent indoor hybrid 3D model. The method allows a wide range of applications from interior navigation to virtual stores.
Author supplied keywords
Cite
CITATION STYLE
Poux, F., Neuville, R., Nys, G. A., & Billen, R. (2018). 3D point cloud semantic modelling: Integrated framework for indoor spaces and furniture. Remote Sensing, 10(9). https://doi.org/10.3390/rs10091412
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.