What can we learn about a scene by watching it for months or years? A video recorded over a long timespan will depict interesting phenomena at multiple timescales, but identifying and viewing them presents a challenge. The video is too long to watch in full, and some things are too slow to experience in real-time, such as glacial retreat or the gradual shift from summer to fall. Timelapse videography is a common approach to summarizing long videos and visualizing slow timescales. However, a timelapse is limited to a single chosen temporal frequency, and often appears flickery due to aliasing. Also, the length of the timelapse video is directly tied to its temporal resolution, which necessitates tradeoffs between those two facets. In this paper, we propose Video Temporal Pyramids, a technique that addresses these limitations and expands the possibilities for visualizing the passage of time. Inspired by spatial image pyramids from computer vision, we developed an algorithm that builds video pyramids in the temporal domain. Each level of a Video Temporal Pyramid visualizes a different timescale; for instance, videos from the monthly timescale are usually good for visualizing seasonal changes, while videos from the one-minute timescale are best for visualizing sunrise or the movement of clouds across the sky. To help explore the different pyramid levels, we also propose a Video Spectrogram to visualize the amount of activity across the entire pyramid, providing a holistic overview of the scene dynamics and the ability to explore and discover phenomena across time and timescales. To demonstrate our approach, we have built Video Temporal Pyramids from ten outdoor scenes, each containing months or years of data. We compare Video Temporal Pyramid layers to naive timelapse and find that our pyramids enable alias-free viewing of longer-term changes. We also demonstrate that the Video Spectrogram facilitates exploration and discovery of phenomena across pyramid levels, by enabling both overview and detail-focused perspectives.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TVCG.2022.3209454DOI Listing

Publication Analysis

Top Keywords

video temporal
20
temporal pyramids
12
video
11
visualizing passage
8
passage time
8
temporal
8
propose video
8
temporal pyramid
8
pyramid levels
8
video spectrogram
8

Similar Publications

Adaptation optimizes sensory encoding for future stimuli.

PLoS Comput Biol

January 2025

Department of Psychology, University of Pennsylvania, Philadelphia, Pennsylvania, United States of America.

Sensory neurons continually adapt their response characteristics according to recent stimulus history. However, it is unclear how such a reactive process can benefit the organism. Here, we test the hypothesis that adaptation actually acts proactively in the sense that it optimally adjusts sensory encoding for future stimuli.

View Article and Find Full Text PDF

Background: Estimates of tick abundance and distribution are used to determine the risk of tick-host contact. Tick surveys provide estimates of distributions and relative abundance for species that remain stationary and wait for passing hosts (i.e.

View Article and Find Full Text PDF

A comprehensive scoping review on machine learning-based fetal echocardiography analysis.

Comput Biol Med

January 2025

Institute of Biomedical Engineering, University of Oxford, Old Road Campus Research Building, Oxford, OX3 7DQ, UK.

Fetal echocardiography (ultrasound of the fetal heart) plays a vital role in identifying heart defects, allowing clinicians to establish prenatal and postnatal management plans. Machine learning-based methods are emerging to support the automation of fetal echocardiographic analysis; this review presents the findings from a literature review in this area. Searches were queried at leading indexing platforms ACM, IEEE Xplore, PubMed, Scopus, and Web of Science, including papers published until July 2023.

View Article and Find Full Text PDF

Artificial neural networks (ANNs) can help camera-based remote photoplethysmography (rPPG) in measuring cardiac activity and physiological signals from facial videos, such as pulse wave, heart rate and respiration rate with better accuracy. However, most existing ANN-based methods require substantial computing resources, which poses challenges for effective deployment on mobile devices. Spiking neural networks (SNNs), on the other hand, hold immense potential for energy-efficient deep learning owing to their binary and event-driven architecture.

View Article and Find Full Text PDF

The Segment Anything model (SAM) is a powerful vision foundation model that is revolutionizing the traditional paradigm of segmentation. Despite this, a reliance on prompting each frame and large computational cost limit its usage in robotically assisted surgery. Applications, such as augmented reality guidance, require little user intervention along with efficient inference to be usable clinically.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!