EVLM: Intent-Driven Edge Vision Language Model for UAV-Based Power Line Inspection

EVLM: Intent-Driven Edge Vision Language Model for UAV-Based Power Line Inspection

2026 IEEE International Conference on Edge Computing and Communications (IEEE EDGE 2026)

Reza Farahani (TU Wien, Austria), Zoha Azimi (AAU, Austria), Ilir Murturi (University of Prishtina, Kosovo), Arda Goknil (SINTEF, Norway), Sagar Sen (SINTEF, Norway), Christian Timmerer (AAU, Austria), Schahram Dustdar (TU Wien, Austria)

Abstract: Inspection of critical infrastructure, such as power lines, is increasingly conducted using unmanned aerial vehicles (UAVs) that capture aerial video for subsequent human review. Although recent edge-based approaches deploy onboard object detectors to identify predefined defect classes, these pipelines remain closed-set, task-specific, and largely decoupled from operator intent and edge resource constraints. This paper introduces EVLM, an intent-driven vision-language framework for onboard UAV-based power line inspection. Given a high-level operator intent, EVLM (i) leverages lightweight histogram-based frame filtering to extract salient key frames under bounded compute budgets, (ii) executes a domain-adapted vision language model (VLM) directly on the UAV for intent-conditioned multimodal reasoning, and (iii) synthesizes structured inspection reports together with a minimal set of evidence frames, replacing continuous raw video transmission with compact semantic outputs. To align the VLM with infrastructure inspection semantics while preserving edge efficiency, we perform parameter-efficient fine-tuning using Low-Rank Adaptation (LoRA), enabling domain specialization without updating the full model parameters. We implement and fully deploy EVLM on an NVIDIA Jetson device representative of UAV-class onboard hardware and evaluate it using 20 publicly released power line inspection video sequences spanning 8 heterogeneous environments and 5 operational intent categories. Experimental results show a data reduction of 94.8 %, with transmitted data decreasing from 485 kB to 25 kB per 4 s segment, corresponding to 72.75 MB versus 3.75 MB over a 10 min inspection mission. EVLM operates feasibly on embedded hardware, maintaining moderate CPU/GPU utilization and bounded power consumption (5.6 W), while producing interpretable, intent-aligned inspection outputs with richer semantic insights than detection-centric baselines.

 

 

 

 

Posted in ATHENA | Comments Off on EVLM: Intent-Driven Edge Vision Language Model for UAV-Based Power Line Inspection

Hadi Amirpour has been elevated to IEEE Senior Member

Assistant Prof. Dr. Hadi Amirpour has been elevated to IEEE Senior Member in recognition of his contributions to multimedia streaming systems.

IEEE Senior Member is the highest professional grade for which an IEEE member can apply. This distinction requires extensive professional experience and demonstrated accomplishments that reflect technical expertise, leadership, and professional maturity. Fewer than 10% of IEEE’s nearly half a million members worldwide have achieved this honor.

 

Posted in ATHENA | Comments Off on Hadi Amirpour has been elevated to IEEE Senior Member

Christian Timmerer Named ACM Senior Member

The Association for Computing Machinery (ACM) has recognized Christian Timmerer as a Senior Member, honoring his professional achievements and contributions to the field of computing.

The ACM Senior Member designation is awarded to individuals who have demonstrated significant performance and commitment within the computing profession. This distinction highlights Christian Timmerer’s ongoing engagement with the research community and his impact on advancing the discipline.

As part of this recognition, he will receive an official ACM Senior Member certificate and pin, and his name will be listed on the ACM Senior Member award page.

Christian Timmerer also expressed his sincere appreciation to colleagues, collaborators, and supporters who contributed throughout the nomination process, emphasizing that this recognition reflects a shared effort within the community.

This honor underscores both his individual accomplishments and his continued dedication to excellence in computing research and practice.

 

Posted in ATHENA, News | Comments Off on Christian Timmerer Named ACM Senior Member

Energy and Compression Efficiency in Large-Scale Video Streaming

Energy and Compression Efficiency in Large-Scale Video Streaming

IEEE International Conference on Image Processing (ICIP 2026)

[PDF]

Mohammad Ghasempour (AAU, Austria), Hadi Amirpour (AAU, Austria),  and Christian Timmerer (AAU, Austria)

Abstract: The rise in large-scale video streaming has led to increased energy demands across the encoding, transmission, and decoding pipeline. While energy consumption in video streaming has been widely studied, encoding decisions are typically made without explicitly accounting for expected content demand. As a result, the impact of view count on energy consumption and compression efficiency remains largely unexplored. This limits the ability to make informed and efficient encoding decisions in real-world streaming scenarios. In this paper, we propose EcoEncode, an analytical framework to evaluate the impact of view count on codec-level encoding decisions and the resulting trade-offs between energy consumption and compression efficiency. We further show that these decisions depend on video content characteristics and encoding configurations. Based on our findings, we provide practical insights to guide the selection of codecs and presets. Experimental results show that view count is a key factor in codec-level decisions. For low-popularity videos, EcoEncode achieves up to 99% energy savings with only 1-4 VMAF points of quality loss. Across all scenarios, the selected configurations lie on or near the Pareto frontier, and EcoEncode improves quality by up to 14 VMAF points over the least energy-consuming configuration.

Posted in ATHENA | Comments Off on Energy and Compression Efficiency in Large-Scale Video Streaming

How Video Streaming Works: Insights from FÄKT (German Video)

As part of the “FÄKT” initiative by the Austrian Academy of Sciences, which produces science videos for audiences aged 10 to 14, a recent episode highlights insights into the world of video streaming: Christian Timmerer, head of the Christian Doppler Laboratory for Adaptive Streaming over HTTP and Emerging Network-based Multimedia Services at the University of Klagenfurt and a two-time Technology & Engineering Emmy® Award recipient, explains how video streaming works on a global scale and how it is continuously optimized.

The short video provides an accessible overview of the technologies behind modern streaming services and demonstrates how research contributes to improving the quality and efficiency of video delivery. Please note that the video is currently available in German only.

Links

Posted in ATHENA | Comments Off on How Video Streaming Works: Insights from FÄKT (German Video)

Long Night of Research 2026 – A Night to Remember

The Lange Nacht der Forschung 2026 (long night of research) turned out to be a truly special evening — one that once again demonstrated how powerful it can be to bring science and research closer to the public. Thanks to the remarkable engagement, creativity, and enthusiasm of everyone involved, complex ideas were transformed into hands-on experiences for a broad and diverse audience.

With more than 9,000 visitors across the Lakeside Science & Technology Park and the University of Klagenfurt campus, the event was a great success. Each individual station contributed to making research tangible, interactive, and inspiring.

Strong Presence of Our Department

Our department was proudly represented with six stations/booths, four of which were hosted by our lab. Together, they showcased cutting-edge research in multimedia, artificial intelligence, and interactive systems, thus demonstrating both scientific depth and real-world impact.

Highlights from Our Lab

At our lab’s four stations, visitors had the opportunity to explore current research in an engaging and interactive way:

Detecting Damage in Wind Turbines with AI (L25)
How can we inspect wind turbines without shutting them down? This station introduced the DORBINE project, where AI-powered drone swarms are used for automated inspection. A two-meter model vividly demonstrated how such intelligent systems could reduce costs and downtime while improving energy efficiency.

Making 3D Video More Realistic (L26)
Visitors were introduced to 3D Gaussian Splatting (3DGS), a next-generation 3D video technology that enables highly realistic rendering of scenes with reduced data requirements. Through hands-on interaction, they experienced how real-world environments can be captured and reproduced as immersive 3D spaces.

Enhancing Video Quality with Super-Resolution (L27)
This station focused on AI-based super-resolution techniques. Attendees could directly compare videos of different quality levels and observe in real time how machine learning reconstructs fine details and textures from low-resolution footage.

Experiencing Multimedia with 3D Interaction (L28)
Using Apple Vision Pro head-mounted displays, visitors explored stereoscopic spatial videos and tested their skills in a 3D dart game. This station highlighted how perception and interaction merge in next-generation multimedia experiences, offering a glimpse into future human-computer interaction.

Making Research Tangible

What made the evening particularly special was not only the technologies themselves but also the way they were communicated: interactive demos, hands-on exploration, and direct conversations with researchers allowed visitors of all ages to engage with science in a meaningful way.

Thank You

A big thank you to everyone who contributed to making this event such a success, through preparation, creativity, and dedication on-site. Events like the Lange Nacht der Forschung thrive on teamwork, and this year was a perfect example.

Posted in ATHENA, News | Comments Off on Long Night of Research 2026 – A Night to Remember

EPS: Efficient Patch Sampling for Video Overfitting in Deep Super-Resolution Model Training

EPS: Efficient Patch Sampling for Video Overfitting
in Deep Super-Resolution Model Training

IEEE Transactions on Circuits and Systems for Video Technology (TCSVT)

[PDF]

Yiying Wei (AAU, Austria), Hadi Amirpour (AAU, Austria),  Jong Hwan Ko (SKKU, South Korea), and Christian Timmerer (AAU, Austria)

Abstract: Leveraging the overfitting property of deep neural networks (DNNs) is trending in video delivery systems to enhance video quality within bandwidth limits. Existing approaches transmit overfitted super-resolution (SR) model streams for low-resolution (LR) bitstreams, which are used to reconstruct high-resolution (HR) videos at the decoder. Although these approaches show promising results, the huge computational costs of training a large number of video frames limit their practical applications. To overcome this challenge, we propose an efficient patch sampling method named EPS for video SR network overfitting, which identifies the most valuable training patches from video frames.
To this end, we first present two low-complexity Discrete Cosine Transform (DCT)-based spatial-temporal features to measure the complexity score of each patch directly. By analyzing the histogram distribution of these features, we then categorize all possible patches into different clusters and select training patches from the cluster with the highest spatial-temporal information. The number of sampled patches is adaptive based on the video content, addressing the trade-off between training complexity and efficiency.
Our method reduces the number of training patches by 75.00% to 91.69%, depending on the resolution and number of clusters, while preserving high video quality and greatly improving training efficiency. Our method speeds up patch sampling by up to 82.1× compared to the state-of-the-art patch sampling technique (EMT).

Posted in ATHENA | Comments Off on EPS: Efficient Patch Sampling for Video Overfitting in Deep Super-Resolution Model Training