In this work, we present a new multi-view depth estimation method NerfingMVS that utilizes both conventional reconstruction and learning-based priors over the recently proposed neural radiance fields (NeRF). Unlike existing neural network based optimization method that relies on estimated correspondences, our method directly optimizes over implicit volumes, eliminating the challenging step of matching pixels in indoor scenes. The key to our approach is to utilize the learning-based priors to guide the optimization process of NeRF. Our system first adapts a monocular depth network over the target scene by finetuning on its MVS reconstruction from COLMAP. Then, we show that the shape-radiance ambiguity of NeRF still exists in indoor environments and propose to address the issue by employing the adapted depth priors to monitor the sampling process of volume rendering. Finally, a per-pixel confidence map acquired by error computation on the rendered image can be used to further improve the depth quality. We further present NerfingMVS++, where a coarse-to-fine depth priors training strategy is proposed to directly utilize sparse SfM points and the uniform sampling is replaced by Gaussian sampling to boost the performance. Experiments show that our NerfingMVS and its extension NerfingMVS++ achieve state-of-the-art performances on indoor datasets ScanNet and NYU Depth V2. In addition, we show that the guided optimization scheme does not sacrifice the original synthesis capability of neural radiance fields, improving the rendering quality on both seen and novel views. Code is available at https://github.com/weiyithu/NerfingMVS.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1109/TPAMI.2023.3263464 | DOI Listing |
Healthc Technol Lett
December 2024
Despite the benefits of minimally invasive surgery, interventions such as laparoscopic liver surgery present unique challenges, like the significant anatomical differences between preoperative images and intraoperative scenes due to pneumoperitoneum, patient pose, and organ manipulation by surgical instruments. To address these challenges, a method for intraoperative three-dimensional reconstruction of the surgical scene, including vessels and tumors, without altering the surgical workflow, is proposed. The technique combines neural radiance field reconstructions from tracked laparoscopic videos with ultrasound three-dimensional compounding.
View Article and Find Full Text PDFSensors (Basel)
December 2024
School of Computer Science, Civil Aviation Flight University of China, Deyang 618307, China.
Accurate classification of three-dimensional (3D) point clouds in real-world environments is often impeded by sensor noise, occlusions, and incomplete data. To overcome these challenges, we propose SMCNet, a robust multimodal framework for 3D point cloud classification. SMCNet combines multi-view projection and neural radiance fields (NeRFs) to generate high-fidelity 2D representations with enhanced texture realism, addressing occlusions and lighting inconsistencies effectively.
View Article and Find Full Text PDFSensors (Basel)
November 2024
Faculty of Industrial Technology, Technical University of Sofia, 1756 Sofia, Bulgaria.
This paper explores the influence of various camera settings on the quality of 3D reconstructions, particularly in indoor crime scene investigations. Utilizing Neural Radiance Fields (NeRF) and Gaussian Splatting for 3D reconstruction, we analyzed the impact of ISO, shutter speed, and aperture settings on the quality of the resulting 3D reconstructions. By conducting controlled experiments in a meeting room setup, we identified optimal settings that minimize noise and artifacts while maximizing detail and brightness.
View Article and Find Full Text PDFPlants (Basel)
November 2024
Hunan Engineering Technology Research Center of Agricultural Rural Informatization, Changsha 410128, China.
Precise acquisition of potted plant traits has great theoretical significance and practical value for variety selection and guiding scientific cultivation practices. Although phenotypic analysis using two dimensional(2D) digital images is simple and efficient, leaf occlusion reduces the available phenotype information. To address the current challenge of acquiring sufficient non-destructive information from living potted plants, we proposed a three dimensional (3D) phenotyping pipeline that combines neural radiation field reconstruction with path analysis.
View Article and Find Full Text PDFJ Imaging Inform Med
December 2024
Department of Radiology, Northwestern Medicine, Northwestern University, Chicago, IL, USA.
Many tasks performed in image-guided procedures can be cast as pose estimation problems, where specific projections are chosen to reach a target in 3D space. In this study, we construct a framework for fluoroscopic pose estimation and compare alternative loss functions and volumetric scene representations. We first develop a differentiable projection (DiffProj) algorithm for the efficient computation of Digitally Reconstructed Radiographs (DRRs) from either Cone-Beam Computerized Tomography (CBCT) or neural scene representations.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!