Christian Timmerer presents at Telecom Seminar Series at TII about HTTP Adaptive Streaming

HTTP Adaptive Streaming – Quo Vadis?

Jun 27, 2024, 04:00 PM Dubai

[Slides]

Abstract: Video traffic on the Internet is constantly growing; networked multimedia applications consume a predominant share of the available Internet bandwidth. A major technical breakthrough and enabler in multimedia systems research and of industrial networked multimedia services certainly was the HTTP Adaptive Streaming (HAS) technique. This resulted in the standardization of MPEG Dynamic Adaptive Streaming over HTTP (MPEG-DASH) which, together with HTTP Live Streaming (HLS), is widely used for multimedia delivery in today’s networks. Existing challenges in multimedia systems research deal with the trade-off between (i) the ever-increasing content complexity, (ii) various requirements with respect to time (most importantly, latency), and (iii) quality of experience (QoE). Optimizing towards one aspect usually negatively impacts at least one of the other two aspects if not both. This situation sets the stage for our research work in the ATHENA Christian Doppler (CD) Laboratory (Adaptive Streaming over HTTP and Emerging Networked Multimedia Services; https://athena.itec.aau.at/), jointly funded by public sources and industry. In this talk, we will present selected novel approaches and research results of the first year of the ATHENA CD Lab’s operation. We will highlight HAS-related research on (i) multimedia content provisioning (machine learning for video encoding); (ii) multimedia content delivery (support of edge processing and virtualized network functions for video networking); (iii) multimedia content consumption and end-to-end aspects (player-triggered segment retransmissions to improve video playout quality); and (iv) novel QoE investigations (adaptive point cloud streaming). We will also put the work into the context of international multimedia systems research.

Biography: Christian Timmerer is a full professor of computer science at Alpen-Adria-Universität Klagenfurt (AAU), Institute of Information Technology (ITEC) and he is the director of the Christian Doppler (CD) Laboratory ATHENA (https://athena.itec.aau.at/). His research interests include multimedia systems, immersive multimedia communication, streaming, adaptation, and quality of experience where he co-authored more than 20 patent applications and more than 300 articles. He was the general chair of WIAMIS 2008, QoMEX 2013, MMSys 2016, and PV 2018 and has participated in several EC-funded projects, notably DANAE, ENTHRONE, P2P-Next, ALICANTE, SocialSensor, COST IC1003 QUALINET, ICoSOLE, and SPIRIT. He also participated in ISO/MPEG work for several years, notably in the area of MPEG-21, MPEG-M, MPEG-V, and MPEG-DASH where he also served as standard editor. In 2012 he cofounded Bitmovin (http://www.bitmovin.com/) to provide professional services around MPEG-DASH where he holds the position of the Chief Innovation Officer (CIO) –- Head of Research and Standardization. Further information at http://timmerer.com.

Posted in ATHENA | Comments Off on Christian Timmerer presents at Telecom Seminar Series at TII about HTTP Adaptive Streaming

Patent Approval for “Per-Title Encoding Using Spatial and Temporal Resolution Downscaling”

Per-Title Encoding Using Spatial and Temporal Resolution Downscaling

US Patent

[PDF]

Hadi Amirpour (Alpen-Adria-Universität Klagenfurt, Austria) and Christian Timmerer (Alpen-Adria-Universität Klagenfurt, Austria)

 

Abstract: Techniques relating to per-title encoding using spatial and temporal resolution downscaling is disclosed. A method for per-title encoding includes receiving a video input comprised of video segments, spatially downscaling the video input, temporally downscaling the video input, encoding the video input to generate an encoded video, then temporally and spatially upscaling the encoded video. Spatially downscaling may include reducing a resolution of the video input, and temporally downscaling may include reducing a framerate of the video input. Objective metrics for the upscaled encoded video show improved quality over conventional methods.

Posted in ATHENA | Comments Off on Patent Approval for “Per-Title Encoding Using Spatial and Temporal Resolution Downscaling”

Samira Afzal talk at 11th Fraunhofer FOKUS MWS

Energy Efficient Video Encoding for Cloud and Edge Computing Instances

11th FOKUS Media Web Symposium

 11th JUN 2024 | Berlin, Germany

 

Abstract: The significant increase in energy consumption within data centers is primarily due to the exponential rise in demand for complex computing workflows and storage resources. Video streaming applications, which are both compute and storage-intensive, account for the majority of today’s internet services. To address this, this talk proposes a novel matching-based method designed to schedule video encoding applications on Cloud resources. The method optimizes for user-defined objectives, including energy consumption, processing time, cost, CO2 emissions, or a a trade-off between these priorities.

Samira Afzal is a postdoctoral researcher at Alpen-Adria-Universität Klagenfurt, Austria, and in collaborating with Bitmovin. Before, she was a postdoctoral researcher at the University of Sao Paulo, researching deeply on IoT, SWARM, and Surveillance Systems. She graduated with her Ph.D. in November 2019 from the University of Campinas (UNICAMP). During her Ph.D., she collaborated with Samsung on a project in the area of mobile video streaming over heterogeneous wireless networks and multipath transmission methods in order to increase perceived
video quality. Further information is available here.

Posted in ATHENA, GAIA | Comments Off on Samira Afzal talk at 11th Fraunhofer FOKUS MWS

ODVista: An Omnidirectional Video Dataset for Super-Resolution and Quality Enhancement Tasks

IEEE International Conference on Image Processing (ICIP 2024)

27-30 October 2024, Abu Dhabi, UAE

[PDF]

Ahmed Telili (TII, UAE), Ibrahim Farhat (TII, UAE), Wassim Hamidouche (TII, UAE), Hadi Amirpour (AAU, Austria)

 

Abstract: Omnidirectional or 360-degree video is being increasingly deployed, largely due to the latest advancements in immersive virtual reality (VR) and extended reality (XR) technology. However, the adoption of these videos in streaming encounters challenges related to bandwidth and latency, particularly in mobility conditions such as with unmanned aerial vehicles (UAVs). Adaptive resolution and compression aim to preserve quality while maintaining low latency under these constraints, yet downscaling and encoding can still degrade quality and introduce artifacts. Machine learning (ML)-based super-resolution (SR) and quality enhancement techniques offer a promising solution by enhancing detail recovery and reducing compression artifacts. However, current publicly available 360-degree video SR datasets lack compression artifacts, which limit research in this field. To bridge this gap, this paper introduces omnidirectional video streaming dataset (ODVista), which comprises 200 high-resolution and high-quality videos downscaled and encoded at four bitrate ranges using the high-efficiency video coding (HEVC)/H.265 standard. Evaluations show that the dataset not only features a wide variety of scenes but also spans different levels of content complexity, which is crucial for robust solutions that perform well in real-world scenarios and generalize across diverse visual environments. Additionally, we evaluate the performance, considering both quality enhancement and runtime, of two handcrafted and two ML-based SR models on the validation and testing sets of ODVista.
Dataset URL: https://github.com/Omnidirectional-video-group/ODVista

 

Posted in ATHENA | Comments Off on ODVista: An Omnidirectional Video Dataset for Super-Resolution and Quality Enhancement Tasks

Long Night of Research/Lange Nacht der Forschung 2024

The Klagenfurt University (AAU) made a strong impression at the 2024 Lange Nacht der Forschung (Long Night of Research) held at Klagenfurt University and Lakeside Park on May 24th, attracting over 8,000 visitors. The Athena Lab, a leading research group within AAU, particularly impressed visitors with its three interactive stations – ATHENA (L20), GAIA (L21) and  SPIRIT (L22) – showcasing its work at the forefront of technology and sustainability.

ATHENA: L20 – How does video work on the Internet?

The ATHENA (L20) station explored the world of video streaming. Visitors learned how content travels from its source to their devices. Through interactive displays, they learned how innovative technologies ensure videos stream quickly and in the best quality possible, reaching Smart TVs seamlessly.

GAIA: L21 – Greener Video Streaming for a Sustainable Future

The GAIA (L21) station aimed to raise visitors’ awareness about the energy consumption and environmental impact of video streaming. It demonstrated how modern technologies and a conscious approach to video streaming can positively impact the environment. This station encouraged visitors to contribute to a greener future.

                                              

SPIRIT: L22 – A look into the future of virtual presence: What do people look like as 3D point clouds?

This station transported visitors to the future of communication. SPIRIT (L22) explored immersive telepresence, where people and objects are no longer confined to 2D video tiles but represented as realistic 3D point clouds for VR/AR glasses. Imagine the iconic “holodeck” from Star Trek coming to life! The station showed what such representations might look like, bridging the gap between the physical and virtual worlds.

                                               

The ATHENA Lab’s stations were a magnet for kids at LNDF! Filled with cool activities and demos, the displays showcasing cutting-edge research kept young minds engaged and curious about the future.

 

Posted in ATHENA, GAIA, SPIRIT | Comments Off on Long Night of Research/Lange Nacht der Forschung 2024

On the Security of Selectively Encrypted HEVC Video Bitstreams

ACM Transactions on Multimedia Computing Communications and Applications (ACM TOMM)

[PDF]

Chen Chen (Tsinghua University, China),  Lingfeng Qu (Guangzhou University, China), Hadi Amirpour (Alpen-Adria-Universität Klagenfurt), Xingjun Wang (Tsinghua University, China),  Christian Timmerer (Alpen-Adria-Universität Klagenfurt), and Zhihong Tian (Guangzhou University, China)

Abstract:

With the growing applications of video, ensuring its security has become of utmost importance. Selective encryption (SE) has gained significant attention in the field of video content protection due to its compatibility with video codecs, favorable visual distortion, and low time complexity. However, few studies consider SE security under cryptographic attacks. To fill this gap, we analyze the security concerns of encrypted bitstreams by SE schemes and propose two known plaintext attacks (KPAs). Then, the corresponding defense is presented against the KPAs.To validate the effectiveness of the KPA, it is applied to attack two existing SE schemes with superior visual degradation in HEVC videos.

Firstly, the encrypted bitstreams are generated using the HEVC encoder with SE (HESE).
Secondly, the video sequences are encoded using H.265/HEVC. During encoding, the selected syntax elements are recorded. Then, the recorded syntax elements are imported into the HEVC decoder using decryption (HDD). By utilizing the encryption parameters and the imported data in the HDD, it becomes possible to reconstruct a significant portion of the original syntax elements before encryption. Finally, the reconstructed syntax elements are compared with the encrypted syntax elements in the HDD, allowing the design of a pseudo-key stream (PKS) through the inverse of the encryption operations. The PKS is used to decrypt the existing SE scheme, and the experimental results provide evidence that the two existing SE schemes are vulnerable to the proposed KPAs.
In the case of single bitstream estimation (SBE), the average correct rate of key stream estimation exceeds 93%. Moreover, with multi-bitstream complementation (MBC), the average estimation accuracy can be further improved to 99%.

 

Posted in ATHENA | Comments Off on On the Security of Selectively Encrypted HEVC Video Bitstreams

Video Encoding Enhancement via Content-Aware Spatial and Temporal Super-Resolution

32nd European Signal Processing Conference (EUSIPCO 2024)

26-30 August 2024, Lyon, France

[PDF]

Yiying Wei (AAU, Austria), Hadi Amirpour (AAU, Austria) Ahmed Telili (INSA Rennes, France), Wassim Hamidouche (INSA Rennes, France), Guo Lu (Shanghai Jiao Tong University, China) and Christian Timmerer (AAU, Austria)

Abstract: Content-aware deep neural networks (DNNs) are trending in Internet video delivery. They enhance quality within bandwidth limits by transmitting videos as low resolution (LR) bitstreams with overfitted super-resolution (SR) model streams to reconstruct high-resolution (HR) video on the decoder end. However, these methods underutilize spatial and temporal re- dundancy, compromising compression efficiency. In response, our proposed video compression framework introduces spatial- temporal video super-resolution (STVSR), which encodes videos into low spatial-temporal resolution (LSTR) content and a model stream, leveraging the combined spatial and temporal reconstruction capabilities of DNNs. Compared to the state-of- the-art approaches that consider only spatial SR, our approach achieves bitrate savings of 18.71% and 17.04% while maintainingthe same PSNR and VMAF, respectively.

Posted in ATHENA | Comments Off on Video Encoding Enhancement via Content-Aware Spatial and Temporal Super-Resolution