Video Super-Resolution Method Using Deformable Convolution-Based Alignment Network.

Sensors (Basel)

Department of Computer Engineering, Dong-A University, Busan 49315, Korea.

Published: November 2022

With the advancement of sensors, image and video processing have developed for use in the visual sensing area. Among them, video super-resolution (VSR) aims to reconstruct high-resolution sequences from low-resolution sequences. To use consecutive contexts within a low-resolution sequence, VSR learns the spatial and temporal characteristics of multiple frames of the low-resolution sequence. As one of the convolutional neural network-based VSR methods, we propose a deformable convolution-based alignment network (DCAN) to generate scaled high-resolution sequences with quadruple the size of the low-resolution sequences. The proposed method consists of a feature extraction block, two different alignment blocks that use deformable convolution, and an up-sampling block. Experimental results show that the proposed DCAN achieved better performances in both the peak signal-to-noise ratio and structural similarity index measure than the compared methods. The proposed DCAN significantly reduces the network complexities, such as the number of network parameters, the total memory, and the inference speed, compared with the latest method.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC9656337PMC
http://dx.doi.org/10.3390/s22218476DOI Listing

Publication Analysis

Top Keywords

video super-resolution
8
deformable convolution-based
8
convolution-based alignment
8
alignment network
8
high-resolution sequences
8
low-resolution sequences
8
low-resolution sequence
8
proposed dcan
8
super-resolution method
4
method deformable
4

Similar Publications

Enhancing space-time video super-resolution via spatial-temporal feature interaction.

Neural Netw

December 2024

College of Electronic and Information Engineering, Tongji University, China; Shanghai Institute of Intelligent Science and Technology, Tongji University, China. Electronic address:

The target of space-time video super-resolution (STVSR) is to increase both the frame rate (also referred to as the temporal resolution) and the spatial resolution of a given video. Recent approaches solve STVSR using end-to-end deep neural networks. A popular solution is to first increase the frame rate of the video; then perform feature refinement among different frame features; and at last, increase the spatial resolutions of these features.

View Article and Find Full Text PDF
Article Synopsis
  • The paper introduces HH-NeRF, a cutting-edge rendering framework that creates high-resolution audio-driven talking portraits quickly and with great detail.
  • It features a detail-aware NeRF module that captures realistic eye movements and fine details while maintaining a rendering speed similar to existing methods.
  • Additionally, it uses a conditional super-resolution module that leverages depth maps and audio features to produce realistic high-resolution videos from low-resolution outputs.
View Article and Find Full Text PDF

FCS videos: Fluorescence correlation spectroscopy in space and time.

Biochim Biophys Acta Gen Subj

November 2024

Centre for BioImaging Sciences, National University of Singapore, 117557 Singapore, Singapore; Institute of Digital Molecular Analytics and Science, 117557 Singapore, Singapore; Department of Chemistry, National University of Singapore, 117543 Singapore, Singapore.

Fluorescence Correlation Spectroscopy (FCS), invented more than 50 years ago is a widely used tool providing information on molecular processes in a variety of samples from materials to life sciences. In the last two decades FCS was multiplexed and ultimately made into an imaging technique that provided maps of molecular parameters over whole sample cross-section. However, it was still limited by a measurement time on the order of minutes.

View Article and Find Full Text PDF

Online video super-resolution (online-VSR) highly relies on an effective alignment module to aggregate temporal information, while the strict latency requirement makes accurate and efficient alignment very challenging. Though much progress has been achieved, most of the existing online-VSR methods estimate the motion fields of each frame separately to perform alignment, which is computationally redundant and ignores the fact that the motion fields of adjacent frames are correlated. In this work, we propose an efficient Temporal Motion Propagation (TMP) method, which leverages the continuity of motion field to achieve fast pixel-level alignment among consecutive frames.

View Article and Find Full Text PDF

Human Tooth Crack Image Analysis with Multiple Deep Learning Approaches.

Ann Biomed Eng

September 2024

Division of Electrical and Computer Engineering, College of Engineering, Louisiana State University, Baton Rouge, LA, 70803, USA.

Tooth cracks, one of the most common dental diseases, can result in the tooth falling apart without prompt treatment; dentists also have difficulty locating cracks, even with X-ray imaging. Indocyanine green (ICG) assisted near-infrared fluorescence (NIRF) dental imaging technique can solve this problem due to the deep penetration of NIR light and the excellent fluorescence characteristics of ICG. This study extracted 593 human cracked tooth images and 601 non-cracked tooth images from NIR imaging videos.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!