arXiv Analytics

Sign in

arXiv:2102.12839 [cs.CV]AbstractReferencesReviewsResources

A deep perceptual metric for 3D point clouds

Maurice Quach, Aladine Chetouani, Giuseppe Valenzise, Frederic Dufaux

Published 2021-02-25Version 1

Point clouds are essential for storage and transmission of 3D content. As they can entail significant volumes of data, point cloud compression is crucial for practical usage. Recently, point cloud geometry compression approaches based on deep neural networks have been explored. In this paper, we evaluate the ability to predict perceptual quality of typical voxel-based loss functions employed to train these networks. We find that the commonly used focal loss and weighted binary cross entropy are poorly correlated with human perception. We thus propose a perceptual loss function for 3D point clouds which outperforms existing loss functions on the ICIP2020 subjective dataset. In addition, we propose a novel truncated distance field voxel grid representation and find that it leads to sparser latent spaces and loss functions that are more correlated with perceived visual quality compared to a binary representation. The source code is available at https://github.com/mauriceqch/2021_pc_perceptual_loss.

Comments: Presented at IS&T Electronic Imaging: Image Quality and System Performance, January 2021
Categories: cs.CV, cs.LG, eess.IV, eess.SP
Related articles: Most relevant | Search more
arXiv:1806.01411 [cs.CV] (Published 2018-06-04)
Learning Scene Flow in 3D Point Clouds
arXiv:1707.02392 [cs.CV] (Published 2017-07-08)
Representation Learning and Adversarial Generation of 3D Point Clouds
arXiv:2011.13005 [cs.CV] (Published 2020-11-25)
PREDATOR: Registration of 3D Point Clouds with Low Overlap