Abstract
A learning-based YOGA (Yet Another Geometry-based Point Cloud Compressor) is proposed. It is flexible, allowing for the separable lossy compression of geometry and color attributes, and variable-rate coding using a single neural model; it is high-efficiency, significantly outperforming the latest G-PCC standard quantitatively and qualitatively, e.g., 25% BD-BR gains using PCQM (Point Cloud Quality Metric) as the distortion assessment, and it is lightweight, e.g., similar runtime as the G-PCC codec, owing to the use of sparse convolution and parallel entropy coding. To this end, YOGA adopts a unified end-to-end learning-based backbone for separate geometry and attribute compression. The backbone uses a two-layer structure, where the downscaled thumbnail point cloud is encoded using G-PCC at the base layer, and upon G-PCC compressed priors, multiscale sparse convolutions are stacked at the enhancement layer to effectively characterize spatial correlations to compactly represent the full-resolution sample. In addition, YOGA integrates the adaptive quantization and entropy model group to enable variable-rate control, as well as adaptive filters for better quality restoration.
Author supplied keywords
Cite
CITATION STYLE
Zhang, J., Chen, T., Ding, D., & Ma, Z. (2023). YOGA: Yet Another Geometry-based Point Cloud Compressor. In MM 2023 - Proceedings of the 31st ACM International Conference on Multimedia (pp. 9070–9081). Association for Computing Machinery, Inc. https://doi.org/10.1145/3581783.3613847
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.