This work presents a variant approach to the monocular SLAM problem focused in exploiting the advantages of a human-robot interaction (HRI) framework. Based upon the delayed inverse-depth feature initialization SLAM (DI-D SLAM), a known monocular technique, several but crucial modifications are introduced taking advantage of data from a secondary monocular sensor, assuming that this second camera is worn by a human. The human explores an unknown environment with the robot, and when their fields of view coincide, the cameras are considered a pseudo-calibrated stereo rig to produce estimations for depth through parallax. These depth estimations are used to solve a related problem with DI-D monocular SLAM, namely, the requirement of a metric scale initialization through known artificial landmarks. The same process is used to improve the performance of the technique when introducing new landmarks into the map. The convenience of the approach taken to the stereo estimation, based on SURF features matching, is discussed. Experimental validation is provided through results from real data with results showing the improvements in terms of more features correctly initialized, with reduced uncertainty, thus reducing scale and orientation drift. Additional discussion in terms of how a real-time implementation could take advantage of this approach is provided.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4029671PMC
http://dx.doi.org/10.3390/s140406317DOI Listing

Publication Analysis

Top Keywords

monocular slam
12
monocular
5
slam autonomous
4
autonomous robots
4
robots enhanced
4
enhanced features
4
features initialization
4
initialization work
4
work presents
4
presents variant
4

Similar Publications

PGMF-VINS: Perpendicular-Based 3D Gaussian-Uniform Mixture Filter.

Sensors (Basel)

October 2024

Academy for Engineering and Technology, Fudan University, Shanghai 200433, China.

Article Synopsis
  • Visual-Inertial SLAM (VI-SLAM) enables low-cost, high-precision applications in areas like robotics, autonomous driving, and AR/VR, focusing on both localization and mapping tasks.
  • Researchers typically emphasize localization while neglecting the robustness of mapping, leading to potential limitations in map quality.
  • The proposed map-point convergence strategy, named PGMF, improves mapping accuracy using perpendicular-based triangulation and a 3D Gaussian-uniform mixture filter, resulting in a significant increase in valid map points and reduced variance compared to existing methods.
View Article and Find Full Text PDF

Monocular Simultaneous Localization and Mapping (SLAM), Visual Odometry (VO), and Structure from Motion (SFM) are techniques that have emerged recently to address the problem of reconstructing objects or environments using monocular cameras. Monocular pure visual techniques have become attractive solutions for 3D reconstruction tasks due to their affordability, lightweight, easy deployment, good outdoor performance, and availability in most handheld devices without requiring additional input devices. In this work, we comprehensively overview the SLAM, VO, and SFM solutions for the 3D reconstruction problem that uses a monocular RGB camera as the only source of information to gather basic knowledge of this ill-posed problem and classify the existing techniques following a taxonomy.

View Article and Find Full Text PDF
Article Synopsis
  • Metric3D v2 is a new geometric foundation model that estimates metric depth and surface normals from single images, crucial for accurate 3D recovery.
  • The model addresses challenges in zero-shot generalization for both normal estimation and depth recovery, with innovative solutions like a camera space transformation module and a joint depth-normal optimization module.
  • Trained on over 16 million images, it outperforms existing methods in multiple benchmarks, offering improved accuracy in recovering 3D structures from diverse internet images.
View Article and Find Full Text PDF

Feature Detection of Non-Cooperative and Rotating Space Objects through Bayesian Optimization.

Sensors (Basel)

July 2024

Department of Mechanical and Aerospace Engineering, Rutgers University, New Brunswick, NJ 08901, USA.

In this paper, we propose a Bayesian Optimization (BO)-based strategy using the Gaussian Process (GP) for feature detection of a known but non-cooperative space object by a chaser with a monocular camera and a single-beam LIDAR in a close-proximity operation. Specifically, the objective of the proposed Space Object Chaser-Resident Assessment Feature Tracking (SOCRAFT) algorithm is to determine the camera directional angles so that the maximum number of features within the camera range is detected while the chaser moves in a predefined orbit around the target. For the chaser-object spatial incentive, rewards are assigned to the chaser states from a combined model with two components: feature detection score and sinusoidal reward.

View Article and Find Full Text PDF

Dense SLAM based on monocular cameras does indeed have immense application value in the field of AR/VR, especially when it is performed on a mobile device. In this paper, we propose a novel method that integrates a light-weight depth completion network into a sparse SLAM system using a multi-basis depth representation, so that dense mapping can be performed online even on a mobile phone. Specifically, we present a specifically optimized multi-basis depth completion network, called BBC-Net, tailored to the characteristics of traditional sparse SLAM systems.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!