In this paper, we introduce an approach for future frames prediction based on a single input image. Our method is able to generate an entire video sequence based on the information contained in the input frame. We adopt an autoregressive approach in our generation process, i.e., the output from each time step is fed as the input to the next step. Unlike other video prediction methods that use "one shot" generation, our method is able to preserve much more details from the input image, while also capturing the critical pixel-level changes between the frames. We overcome the problem of generation quality degradation by introducing a "complementary mask" module in our architecture, and we show that this allows the model to only focus on the generation of the pixels that need to be changed, and to reuse those that should remain static from its previous frame. We empirically validate our methods against various video prediction models on the UT Dallas Dataset, and show that our approach is able to generate high quality realistic video sequences from one static input image. In addition, we also validate the robustness of our method by testing a pre-trained model on the unseen ADFES facial expression dataset. We also provide qualitative results of our model tested on a human action dataset: The Weizmann Action database.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC9099507 | PMC |
http://dx.doi.org/10.3390/s22093533 | DOI Listing |
BMC Res Notes
December 2024
Department of Computer Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran.
This dataset contains demographic, morphological and pathological data, endoscopic images and videos of 191 patients with colorectal polyps. Morphological data is included based on the latest international gastroenterology classification references such as Paris, Pit and JNET classification. Pathological data includes the diagnosis of the polyps including Tubular, Villous, Tubulovillous, Hyperplastic, Serrated, Inflammatory and Adenocarcinoma with Dysplasia Grade & Differentiation.
View Article and Find Full Text PDFSci Rep
December 2024
School of Psychological Science, The University of Western Australia, 35 Stirling Highway, Perth, WA, 6009, Australia.
Investigations into whether playing action video games (AVGs) benefit other tasks, such as driving, have traditionally focused on gaming experience (i.e., hours played).
View Article and Find Full Text PDFJ Clin Gastroenterol
October 2024
Division of Gastroenterology and Hepatology, Inflammatory Bowel Disease Center, Mayo Clinic, Jacksonville, FL.
Background: Video capsule retention is a complication that can have serious consequences in patients with Crohn's disease (CD). The patency capsule was developed to detect small bowel strictures. The usefulness of patency capsules in patients who do not have evidence of small bowel disease on imaging is uncertain.
View Article and Find Full Text PDFVet Sci
December 2024
Center for Animals and Public Policy, Cummings School of Veterinary Medicine, Tufts University, 200 Westboro Rd., North Grafton, MA 01536, USA.
Youth mental health interventions incorporating trained therapy animals are increasingly popular, but more research is needed to understand the specific interactive behaviors between participants and therapy dogs. Understanding the role of these interactive behaviors is important for supporting both intervention efficacy and animal welfare and well-being. The goal of this study was to develop ethograms to assess interactive behaviors (including both affiliative and stress-related behaviors) of participants and therapy dogs during a social stress task, explore the relationship between human and dog behaviors, and assess how these behaviors may vary between experimental conditions with varying levels of physical contact with the therapy dog.
View Article and Find Full Text PDFFront Neurol
December 2024
Institut de Recherche Oto-Neurologique (IRON), Paris, France.
Introduction: While most head movements in daily life are active, most tools used to assess vestibular deficits rely on passive head movements. A single gain value is not sufficient to quantify gaze stabilization efficiency during active movements in vestibular deficit patients. Moreover, during active gaze shifts, anticipatory mechanisms come into play.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!