Eye-Tracking, Quality Assessment, and QoE Prediction Models for Point Cloud Videos: Extended Analysis of the ComPEQ-MR Dataset

Eye-Tracking, Quality Assessment, and QoE Prediction Models for Point Cloud Videos: Extended Analysis of the ComPEQ-MR Dataset

IEEE Access 2025

[PDF]

Shivi Vats (AAU, Austria), Minh Nguyen (AAU, Austria)*, Christian Timmerer (AAU, Austria), Hermann Hellwagner (AAU, Austria)

Abstract: Point cloud videos, also termed dynamic point clouds (DPCs), have the potential to provide immersive experiences with six degrees of freedom (6DoF). However, there are still several open issues in understanding the Quality of Experience (QoE) and visual attention of end users while experiencing 6DoF volumetric videos. For instance, the quality impact of compressing DPCs, which requires a significant amount of both time and computational resources, needs further investigation. Also, QoE prediction models for DPCs in 6DoF have rarely been developed due to the lack of visual quality databases. Furthermore, visual attention in 6DoF is hardly explored, which impedes research into more sophisticated approaches for adaptive streaming of DPCs. In this paper, we review and analyze in detail the open-source Compressed Point cloud dataset with Eye-tracking and Quality assessment in Mixed Reality (ComPEQ–MR). The dataset, initially presented in [24], comprises 4 uncompressed (raw) DPCs as well as compressed versions processed by Moving Picture Experts Group (MPEG) reference tools (i.e., VPCC and 2 GPCC variants). The dataset includes eye-tracking data of 41 study participants watching the raw DPCs with 6DoF, yielding 164 visual attention maps. We analyze this data and present head and gaze movement results here. The dataset also includes results from subjective tests conducted to assess the quality of the DPCs, each both uncompressed and compressed with 12 levels of distortion, resulting in 2132 quality scores. This work presents the QoE performance results of the compression techniques, the factors with significant impact on participant ratings, and the correlation of the objective Peak Signal-to-Noise Ratio (PSNR) metrics with Mean Opinion Scores (MOS). The results indicate superior performance of the VPCC codec as well as significant variations in quality ratings based on codec choice, bitrate, and quality/distortion level, providing insights for optimizing point cloud video compression in MR applications. Finally, making use of the subjective scores, we trained and evaluated models for QoE prediction for DPCs compressed using the pertinent MPEG tools. We present the models and their prediction results, noting that the fine-tuned ITU-T P.1203 models exhibit good correlation with the subjective ratings. The dataset is available at https://ftp.itec.aau.at/datasets/ComPEQ-MR/.

* Minh Nguyen is currently a Research Associate at Fraunhofer FOKUS, Germany but this work was done when he was working for AAU.
Posted in SPIRIT | Comments Off on Eye-Tracking, Quality Assessment, and QoE Prediction Models for Point Cloud Videos: Extended Analysis of the ComPEQ-MR Dataset

STEP-MR: A Subjective Testing and Eye-Tracking Platform for Dynamic Point Clouds in Mixed Reality

STEP-MR: A Subjective Testing and Eye-Tracking Platform for Dynamic Point Clouds in Mixed Reality

MMM 2026

January 29 – January 31, 2026

Prague, Czech Republic

[PDF, Poster]

Shivi Vats (AAU, Austria), Christian Timmerer (AAU, Austria), Hermann Hellwagner (AAU, Austria)

Abstract: The use of point cloud (PC) streaming in mixed reality (MR) environments is of particular interest due to the immersiveness and the six degrees of freedom (6DoF) provided by the 3D content. However, this immersiveness requires significant bandwidth. Innovative solutions have been developed to address these challenges, such as PC compression and/or spatially tiling the PC to stream different portions at different quality levels. This paper presents a brief overview of a Subjective Testing and Eye-tracking Platform for dynamic point clouds in Mixed Reality (STEP-MR) for the Microsoft HoloLens 2. STEP-MR was used to conduct subjective tests (described in [1]) with 41 participants, yielding over 2000 responses and more than 150 visual attention maps, the results of which can be used, among other things, to improve dynamic (animated) point cloud streaming solutions mentioned above. Building on our previous platform, the new version now enables eye-tracking tests, including calibration and heatmap generation. Additionally, STEP-MR features modifications to the subjective tests’ functionality, such as a new rating scale and adaptability to participant movement during the tests, along with other user experience changes.

[1] Nguyen, M., Vats, S., Zhou, X., Viola, I., Cesar, P., Timmerer, C., & Hellwagner, H. (2024). ComPEQ-MR: Compressed Point Cloud Dataset with Eye Tracking and Quality Assessment in Mixed Reality. Proceedings of the 15th ACM Multimedia Systems Conference, 367–373. https://doi.org/10.1145/3625468.3652182
Posted in SPIRIT | Comments Off on STEP-MR: A Subjective Testing and Eye-Tracking Platform for Dynamic Point Clouds in Mixed Reality

SEED: Energy and Emission Estimation Dataset for Adaptive Video Streaming

SEED: Energy and Emission Estimation Dataset for Adaptive Video Streaming

IEEE VCIP 2025

December 1 – December 4, 2025

Klagenfurt, Austria

[PDF]

Samira Afzal (Baylor University), Narges Mehran (Salzburg Research Forschungsgesellschaft mbH), Farzad Tashtarian (AAU, Austria), Andrew C. Freeman (Baylor University), Radu Prodan (University of Innsbruck), Christian Timmerer (AAU, Austria)

Abstract: The environmental impact of video streaming is gaining more attention due to its growing share in global internet traffic and energy consumption. To support accurate and transparent sustainability assessments, we present SEED (Streaming Energy and Emission Dataset)}: an open dataset for estimating energy usage and CO2 emissions in adaptive video streaming. SEED comprises over 500 video segments. It provides segment-level measurements of energy consumption and emissions for two primary stages: provisioning, which encompasses encoding and storage on cloud infrastructure, and end-user consumption, including network interface retrieval, video decoding, and display on end-user devices. The dataset covers multiple codecs (AVC, HEVC), resolutions, bitrates, cloud instance types, and geographic regions, reflecting real-world variations in computing efficiency and regional carbon intensity. By combining empirical benchmarks with component-level energy models, \dataset{} enables detailed analysis and supports the development of energy- and emission-aware adaptive bitrate (ABR) algorithms. The dataset is publicly available at: https://github.com/cd-athena/SEED.

SEED is available at: https://github.com/cd-athena/SEED

Posted in ATHENA | Comments Off on SEED: Energy and Emission Estimation Dataset for Adaptive Video Streaming

Resolution vs Quantization: A Trade-Off for Hologram Compression

Resolution vs Quantization: A Trade-Off for Hologram Compression

Ayman Alkhateeb, Hadi Amirpour,Christian Timmerer

Alpen-Adria-Universität , Klagenfurt, Austria

PCS 2025  Achen , German

Holographic imaging offers a path to true three-dimensional visualization for applications such as augmented andvirtual reality, but the immense data size of high-quality holo-grams prevents their practical adoption. This paper investigatesa pre-processing strategy to improve the compression of holo-graphic data using the High-Efficiency Video Coding (HEVC) standard. By downsampling the hologram before encoding andsubsequently upsampling it after decoding, we demonstrate thatit is possible to achieve better reconstruction quality at lowbitrates compared to encoding the full-resolution data. Thiscounterintuitive result basically comes from the reduction inspatial complexity, which allows the HEVC encoder to allocate more bits to preserving critical high-frequency information that would otherwise be lost. Although the hologram phase is highly sensitive to scaling,the overall perceptual quality improves at bitrates below 1 Bpp, with gains of approximately 0.1 in SSIM and 0.015 in VIF. Our work underscores a critical principle in holographic codec design: optimizing the trade-off between spatial complexity and quantization error is key to maximizing reconstruction quality, especially in bandwidth-constrained environments.

Posted in HoloSense | Comments Off on Resolution vs Quantization: A Trade-Off for Hologram Compression

Multiview x265 vs. MV-HEVC: An Efficiency Analysis for Stereoscopic Videos

Multiview x265 vs. MV-HEVC:
An Efficiency Analysis for Stereoscopic Videos

PCS 2025

December 8 – December 11, 2025

Aachen, Germany

[PDF]

Kamran Qureshi, Hadi Amirpour, Christian Timmerer

Abstract: With the increasing demand for immersive video experiences, efficient compression of multiview content has become crucial for reducing storage and transmission costs. The introduction of stereoscopic video support on head-mounted displays, along with the emergence of smartphones capable of easily capturing stereoscopic videos, further highlights the need for optimized encoding solutions. Although the efficient but computationally intensive Multi-View High Efficiency Video Coding (MV-HEVC) standard has been available since 2014, only recently has x265—a real-time open-source HEVC encoder—introduced support for multiview encoding. This work (i) evaluates the encoding efficiency of multiview x265 across all presets, and compares it with MV-HEVC, (ii) proposes a perceptual quality–aware preset selection method, and (iii) conducts a comparative study on single-view and stereoscopic videos.

Posted in ATHENA | Comments Off on Multiview x265 vs. MV-HEVC: An Efficiency Analysis for Stereoscopic Videos

NeVES: Real-Time Neural Video Enhancement for HTTP Adaptive Streaming

NeVES: Real-Time Neural Video Enhancement for HTTP Adaptive Streaming

IEEE VCIP 2025

December 1 – December 4, 2025

Klagenfurt, Austria

[PDF]

Daniele Lorenzi, Farzad Tashtarian, Christian Timmerer

Abstract: Enhancing low-quality video content is a task that has raised particular interest since recent developments in deep learning. Since most of the video content consumed worldwide is delivered over the Internet via HTTP Adaptive Streaming (HAS), implementing these techniques on web browsers would ease the access to visually-enhanced content on user devices.

In this paper, we present NeVES, a multimedia system capable of enhancing the quality of video content streamed through HAS in real time.

The demo is available at: https://github.com/cd-athena/NeVES.

Posted in ATHENA | Comments Off on NeVES: Real-Time Neural Video Enhancement for HTTP Adaptive Streaming

Perceptual Quality Assessment of Spatial Videos on Apple Vision Pro

Perceptual Quality Assessment of Spatial Videos on Apple Vision Pro

ACMMM IXR 2025

October 27 – October 31, 2025

Dublin, Ireland

[PDF]

Afshin Gholami, Sara Baldoni, Federica Battisti, Wei Zhou, Christian Timmerer, Hadi Amirpour

Abstract: Immersive stereoscopic/3D video experiences have entered a new era with the advent of smartphones capable of capturing spatial videos, advanced video codecs optimized for multiview content, and Head Mounted Displays (HMD s) that natively support spatial video playback. In this work, we evaluate the quality of spatial videos encoded using optimized x265 software implementations of MV-HEVC on the AVP and compare them with their corresponding 2D versions through a subjective test.

To support this study, we introduce SV-QoE, a novel dataset comprising video clips rendered with a twin-camera setup that replicates the human inter-pupillary distance. Our analysis reveals that spatial videos consistently deliver a superior Quality of Experience ( QoE ) when encoded at identical bitrates, with the benefits becoming more pronounced at higher bitrates. Additionally, renderings at closer distances exhibit significantly enhanced video quality and depth perception, highlighting the impact of spatial proximity on immersive viewing experiences.

We further analyze the impact of disparity on depth perception and examine the correlation between Mean Opinion Score (MOS ) and established objective quality metrics such as PSNR, SSIM, MS-SSIM, VMAF, and AVQT. Additionally, we explore how video quality and depth perception together influence overall quality judgments.

 

Posted in ATHENA | Comments Off on Perceptual Quality Assessment of Spatial Videos on Apple Vision Pro