Planar digital image sensors facilitate broad applications in a wide range of areas, and the number of pixels has scaled up rapidly in recent years. However, the practical performance of imaging systems is fundamentally limited by spatially nonuniform optical aberrations originating from imperfect lenses or environmental disturbances. Here we propose an integrated scanning light-field imaging sensor, termed a meta-imaging sensor, to achieve high-speed aberration-corrected three-dimensional photography for universal applications without additional hardware modifications. Instead of directly detecting a two-dimensional intensity projection, the meta-imaging sensor captures extra-fine four-dimensional light-field distributions through a vibrating coded microlens array, enabling flexible and precise synthesis of complex-field-modulated images in post-processing. Using the sensor, we achieve high-performance photography up to a gigapixel with a single spherical lens without a data prior, leading to orders-of-magnitude reductions in system capacity and costs for optical imaging. Even in the presence of dynamic atmosphere turbulence, the meta-imaging sensor enables multisite aberration correction across 1,000 arcseconds on an 80-centimetre ground-based telescope without reducing the acquisition speed, paving the way for high-resolution synoptic sky surveys. Moreover, high-density accurate depth maps can be retrieved simultaneously, facilitating diverse applications from autonomous driving to industrial inspections.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC9712118 | PMC |
http://dx.doi.org/10.1038/s41586-022-05306-8 | DOI Listing |
Nat Commun
January 2025
Department of Electronic Engineering, Tsinghua University, Beijing, China.
Optical neural networks are considered next-generation physical implementations of artificial neural networks, but their capabilities are limited by on-chip integration scale and requirement for coherent light sources. This study proposes a spectral convolutional neural network (SCNN) with matter meta-imaging. The optical convolutional layer is implemented by integrating very large-scale and pixel-aligned spectral filters on CMOS image sensor.
View Article and Find Full Text PDFLight Sci Appl
January 2025
Innovation Center for Advanced Medical Imaging and Intelligent Medicine, Guangzhou Institute of Technology, Xidian University, Guangzhou, 510555, Guangdong, China.
A novel monocular depth-sensing camera based on meta-imaging sensor technology has been developed, offering more precise depth sensing with millimeter-level accuracy and enhanced robustness compared to conventional 2D and light-field cameras.
View Article and Find Full Text PDFLight Sci Appl
September 2024
Department of Automation, Tsinghua University, Beijing, 100084, China.
Nature
December 2022
Department of Automation, Tsinghua University, Beijing, China.
Planar digital image sensors facilitate broad applications in a wide range of areas, and the number of pixels has scaled up rapidly in recent years. However, the practical performance of imaging systems is fundamentally limited by spatially nonuniform optical aberrations originating from imperfect lenses or environmental disturbances. Here we propose an integrated scanning light-field imaging sensor, termed a meta-imaging sensor, to achieve high-speed aberration-corrected three-dimensional photography for universal applications without additional hardware modifications.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!