Understanding brain response to audiovisual stimuli is a key challenge in understanding neuronal processes. In this paper, we describe our effort aimed at reconstructing video frames from observed functional MRI images. We also demonstrate that our model can predict visual objects. Our method constructs an autoencoder model for a set of training video segments to code video streams into their corresponding latent representations. Next, we learn a mapping from the observed fMRI response to the corresponding latent video frame representation. Finally, we pass the latent vectors computed using the fMRI response through the decoder to reconstruct the predicted image. We show that the representations of video frames and those constructed from corresponding fMRI images are highly clustered, the latent representations can be used to predict objects in video frames using just the fMRI frames, and fMRI responses can be used to reconstruct the inputs to predict the presence of faces.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC10831884PMC
http://dx.doi.org/10.1016/j.isci.2024.108819DOI Listing

Publication Analysis

Top Keywords

video frames
12
corresponding latent
8
latent representations
8
fmri response
8
frames fmri
8
video
6
fmri
5
see? predicting
4
predicting visual
4
visual features
4

Similar Publications

Gastrointestinal tract-related cancers pose a significant health burden, with high mortality rates. In order to detect the anomalies of the gastrointestinal tract that may progress to cancer, a video capsule endoscopy procedure is employed. The number of video capsule endoscopic ( ) images produced per examination is enormous, which necessitates hours of analysis by clinicians.

View Article and Find Full Text PDF

High-Definition, Video-Rate Triple-Channel NIR-II Imaging Using Shadowless Lamp Excitation and Illumination.

ACS Nano

January 2025

State Key Laboratory of Extreme Photonics and Instrumentations, Centre for Optical and Electromagnetic Research, College of Optical Science and Engineering, International Research Center for Advanced Photonics, Zhejiang University, Hangzhou 310058, China.

Multichannel imaging in the second near-infrared (NIR-II) window offers vital and comprehensive information for complex surgical environments, yet a simple, high-quality, video-rate multichannel imaging method with low safety risk remains to be proposed. Centered at the superior NIR-IIx window of 1400-1500 nm, triple-channel imaging coordinated with 1000-1100 and 1700-1880 nm (NIR-IIc) achieves exceptional clarity and an impressive signal-to-crosstalk ratio as high as 22.10.

View Article and Find Full Text PDF

Key frame extraction algorithm for surveillance videos using an evolutionary approach.

Sci Rep

January 2025

Department of Computer Science and Engineering, Amrita School of Computing, Coimbatore, Amrita Vishwa Vidyapeetham, India, 641112.

With rapid technological advancements, videos are captured, stored, and shared in multiple formats, increasing the requirement for summarization techniques to enable shorter viewing durations. Key Frame Extraction (KFE) algorithms are crucial in video summarization, compression, and offline analysis. This study aims to develop an efficient KFE approach for generic videos.

View Article and Find Full Text PDF

Neuromorphic-enabled video-activated cell sorting.

Nat Commun

December 2024

State Key Laboratory of Precision Measurement Technology and Instrument, Department of Precision Instrument, Tsinghua University, Beijing, China.

Imaging flow cytometry allows image-activated cell sorting (IACS) with enhanced feature dimensions in cellular morphology, structure, and composition. However, existing IACS frameworks suffer from the challenges of 3D information loss and processing latency dilemma in real-time sorting operation. Herein, we establish a neuromorphic-enabled video-activated cell sorter (NEVACS) framework, designed to achieve high-dimensional spatiotemporal characterization content alongside high-throughput sorting of particles in wide field of view.

View Article and Find Full Text PDF

FPANet: Frequency-based video demoiréing using frame-level post alignment.

Neural Netw

December 2024

Department of Artificial Intelligence, Korea University, South Korea. Electronic address:

Moiré patterns, created by the interference between overlapping grid patterns in the pixel space, degrade the visual quality of images and videos. Therefore, removing such patterns (demoiréing) is crucial, yet remains a challenge due to their complexities in sizes and distortions. Conventional methods mainly tackle this task by only exploiting the spatial domain of the input images, limiting their capabilities in removing large-scale moiré patterns.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!