AI Article Synopsis

  • The interest in image fusion, which combines images from different sources like infrared and visible light, has spurred the development of better assessment methods for these fused images.
  • Previous research using subjective ratings and computational metrics often did not align with actual human performance in tasks utilizing these fused images.
  • The current study introduced eye-tracking to measure how well participants could follow a camouflaged person in various fused video formats, revealing that fused displays generally improved tracking accuracy compared to raw inputs, highlighting the potential for these new assessment methods.

Article Abstract

The increased interest in image fusion (combining images of two or more modalities such as infrared and visible light radiation) has led to a need for accurate and reliable image assessment methods. Previous work has often relied upon subjective quality ratings combined with some form of computational metric analysis. However, we have shown in previous work that such methods do not correlate well with how people perform in actual tasks utilising fused images. The current study presents the novel use of an eye-tracking paradigm to record how accurately participants could track an individual in various fused video displays. Participants were asked to track a man in camouflage outfit in various input videos (visible and infrared originals, a fused average of the inputs; and two different wavelet-based fused videos) whilst also carrying out a secondary button-press task. The results were analysed in two ways, once calculating accuracy across the whole video, and by dividing the video into three time sections based on video content. Although the pattern of results depends on the analysis, the accuracy for the inputs was generally found to be significantly worse than that for the fused displays. In conclusion, both approaches have good potential as new fused video assessment methods, depending on what task is carried out.

Download full-text PDF

Source
http://dx.doi.org/10.1163/156856807781503659DOI Listing

Publication Analysis

Top Keywords

fused videos
8
assessment methods
8
previous work
8
fused video
8
fused
6
video
5
assessment fused
4
videos scanpaths
4
scanpaths comparison
4
comparison data
4

Similar Publications

It is important in the rising demands to have efficient anomaly detection in camera surveillance systems for improving public safety in a complex environment. Most of the available methods usually fail to capture the long-term temporal dependencies and spatial correlations, especially in dynamic multi-camera settings. Also, many traditional methods rely heavily on large labeled datasets, generalizing poorly when encountering unseen anomalies in the process.

View Article and Find Full Text PDF

As digital media grows, there is an increasing demand for engaging content that can captivate audiences. Along with that, the monetary conversion of those engaging videos is also increased. This leads to the way for more content-driven videos, which can generate revenue.

View Article and Find Full Text PDF

Adolescent idiopathic scoliosis (AIS), which typically occurs in patients between the ages of 10 and 18, can be caused by a variety of reasons, and no definitive cause has been found. Early diagnosis of AIS or timely recognition of progression is crucial for the prevention of spinal deformity and the reduction of the risk of surgery or postponement. However, it remains a significant challenge.

View Article and Find Full Text PDF

This Letter reports what we believe to be a novel schlieren approach with adaptive temporal resolution. The fundamental concept of this approach is to fuse an event-based camera and a low-speed frame-based camera to generate high-frame-rate videos by leveraging the strengths of both. Using a novel experimental setup, events and frames are accurately aligned in both space and time.

View Article and Find Full Text PDF

The purpose of this study is to investigate how deep learning and other artificial intelligence (AI) technologies can be used to enhance the intelligent level of dance instruction. The study develops a dance action recognition and feedback model based on the Graph Attention Mechanism (GA) and Bidirectional Gated Recurrent Unit (3D-Resnet-BigRu). In this model, time series features are captured using BiGRU after 3D-ResNet is inserted to extract video features.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!