Characterization of the Quality of Experience and Immersion of Point Cloud Videos in Augmented Reality Through a Subjective Study
Tóm tắt
Point cloud streaming has recently attracted research attention as it has the potential to provide six degrees of freedom movement, which is essential for truly immersive media. The transmission of point clouds requires high-bandwidth connections, and adaptive streaming is a promising solution to cope with fluctuating bandwidth conditions. Thus, understanding the impact of different factors in adaptive streaming on the Quality of Experience (QoE) becomes fundamental. Point clouds have been evaluated in Virtual Reality (VR), where viewers are completely immersed in a virtual environment. Augmented Reality (AR) is a novel technology and has recently become popular, yet quality evaluations of point clouds in AR environments are still limited to static images. In this paper, we perform a subjective study of four impact factors on the QoE of point cloud video sequences in AR conditions, including encoding parameters (quantization parameters, QPs), quality switches, viewing distance, and content characteristics. The experimental results show that these factors significantly impact the QoE. The QoE decreases if the sequence is encoded at high QPs and/or switches to lower quality and/or is viewed at a shorter distance, and vice versa. Additionally, the results indicate that the end user is not able to distinguish the quality differences between two quality levels at a specific (high) viewing distance. An intermediate-quality point cloud encoded at geometry QP (G-QP) 24 and texture QP (T-QP) 32 and viewed at 2.5m can have a QoE (i.e., score 6.5 out of 10) comparable to a high-quality point cloud encoded at 16 and 22 for G-QP and T-QP, respectively, and viewed at a distance of 5m. Regarding content characteristics, objects with lower contrast can yield better quality scores. Participants’ responses reveal that the visual quality of point clouds has not yet reached an immersion level as desired. The average QoE of the highest visual quality is less than 8 out of 10. There is also a good correlation between objective metrics (e.g., color Peak Signal-to-Noise Ratio (PSNR) and geometry PSNR) and the QoE score. Especially the Pearson correlation coefficients of color PSNR is 0.84. Finally, we found that machine learning models are able to accurately predict the QoE of point clouds in AR environments. The subjective test results and questionnaire responses are available on GitHub: https://github.com/minhkstn/QoE-and-Immersion-of-Dynamic-Point-Cloud.
Từ khóa
#Point clouds #quality of experience #subjective tests #augmented realityTài liệu tham khảo
10.1007/s10055-021-00513-6
10.1109/ACCESS.2020.3024633
10.1109/TCSVT.2021.3101484
10.1145/3599184.3599188
10.1109/TENCON.2000.888774
10.1109/QoMEX58391.2023.10178579
10.1109/TMM.2022.3148585
10.1111/j.1475-1313.1997.tb00072.x
10.1016/j.patcog.2015.03.009
10.1109/COMST.2018.2862938
whaley, 2005, The interquartile range: Theory and estimation
10.1109/TIP.2016.2575005
10.1145/2910017.2910597
d’eon, 2017, 8i Voxelized Full Bodies version 2-A Voxelized Point Cloud Dataset
10.1109/ACCESS.2020.3004125
2023, Microsoft Hololens 2
10.1109/QoMEX58391.2023.10178443
10.1109/COMST.2023.3263252
10.1109/ICIP46576.2022.9897602
10.1145/3343031.3350917
10.1109/QoMEX48832.2020.9123081
10.1145/3343031.3350543
10.1109/ACCESS.2021.3107619
10.1109/QoMEX58391.2023.10178610
10.1109/MMSP.2017.8122249
vats, 2023, A Platform for Subjective Quality Assessment in Mixed Reality Environments
10.1109/ACCESS.2020.3015556
10.1109/VCIP53242.2021.9675431
10.1145/3145534
2022, Video Point Cloud Compression—VPCC—MPEG-PCC-TMC2 Test Model Candidate Software
2017
10.1109/QoMEX48832.2020.9123121
abdi, 2010, Tukey’s honestly significant difference (HSD) test, Encyclopedia Res Design, 3, 1
takahashi, 2021, Pcx—Point Cloud Importer/Renderer for Unity
10.1109/JETCAS.2018.2885981
10.1007/s10055-020-00446-6
10.1016/0169-7439(89)80095-4
10.1016/j.ecns.2019.09.006
pedregosa, 2012, Scikit-learn: Machine learning in Python, J Mach Learn Res, 12, 2825
10.1109/JETCAS.2019.2898622
10.3390/s22134915
perry, 2021
2023, Scalable Platform for Innovations on Real-Time Immersive Telepresence (SPIRIT) Project Funded by the European Union Grant Agreement 101070672
2023, ICIP 2020
10.1109/TCSVT.2012.2221191
10.1145/1943552.1943572
10.4097/kjae.2015.68.6.540
10.1109/ACCESS.2022.3198995
10.1109/ACCESS.2022.3149592
10.3389/frvir.2023.1130864
10.1017/ATSIP.2020.12
2023, Methodologies for the subjective assessment of the quality of television images
10.1109/TVCG.2007.70441
2023, MPEG-3DG MPEG-PCC-Dmetric Metric Software
krivoku?a, 2018, A volumetric approach to point cloud compression, arXiv 1810 00484
10.1145/3386292.3397117