M3-Unity: Deciphering Perceptual Quality in Colored Point Cloud: Prioritizing Geometry or Texture Distortion?
Point clouds represent one of the prevalent formats for 3D content. Distortions introduced at various stages in the point cloud processing pipeline affect the visual quality, altering their geometric composition, texture information, or both. Understanding and quantifying the impact of the distortion domain on visual quality is vital for driving rate optimization and guiding post-processing steps to improve the quality of experience. In this paper, we propose a multi-task guided multi-modality no reference metric (M3-Unity), which utilizes 4 types of modalities across attributes and dimensionalities to represent point clouds. An attention mechanism establishes inter/intra associations among 3D/2D patches, which can complement each other, yielding local and global features, to fit the highly nonlinear property of the human vision system. A multi-task decoder involving distortion-type classification selects the best association among 4 modalities, aiding the regression task and enabling the in-depth analysis of the interplay between geometrical and textural distortions. Furthermore, our framework design and attention strategy enable us to measure the impact of individual attributes and their combinations, providing insights into how these associations contribute particularly in relation to distortion type. Extensive experimental results on 4 datasets consistently outperform the state-of-the-art metrics by a large margin.
Download
An implementation of this metric can be downloaded by accessing here.
License
Copyright (c) Xuemei Zhou, Irene Viola, Yunlu Chen, Jiahuan Pei, Pablo Cesar. Centrum Wiskunde & Informatica (CWI).
Reference
If you wish to use any of the provided material, we kindly ask you to cite:
-
Visual-Saliency Guided Multi-modal Learning for No Reference Point Cloud Quality Assessment.
In 3rd Workshop on QoEVMA, ACM Multimedia 2024,
2024.
