Depth estimation is a classical problem in computer vision, which typically relies on either a depth sensor or stereo matching alone. The depth sensor provides real-time estimates in repetitive and textureless regions where stereo matching is not effective. However, stereo matching can obtain more accurate results in rich texture regions and object boundaries where the depth sensor often fails. We fuse stereo matching and the depth sensor using their complementary characteristics to improve the depth estimation. Here, texture information is incorporated as a constraint to restrict the pixel's scope of potential disparities and to reduce noise in repetitive and textureless regions. Furthermore, a novel pseudo-two-layer model is used to represent the relationship between disparities in different pixels and segments. It is more robust to luminance variation by treating information obtained from a depth sensor as prior knowledge. Segmentation is viewed as a soft constraint to reduce ambiguities caused by under- or over-segmentation. Compared to the average error rate 3.27% of the previous state-of-the-art methods, our method provides an average error rate of 2.61% on the Middlebury datasets, which shows that our method performs almost 20% better than other "fused" algorithms in the aspect of precision.
Download full-text PDF |
Source |
---|---|
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4570453 | PMC |
http://dx.doi.org/10.3390/s150820894 | DOI Listing |
Sci Rep
January 2025
The Key Laboratory for Agricultural Machinery Intelligent Control and Manufacturing of Fujian Education Institutions, Wuyi University, Nanping, 354300, Fujian, China.
This paper proposes an adaptive real-time tillage depth control system for electric rotary tillers, based on Linear Active Disturbance Rejection Control (LADRC), to improve tillage depth accuracy in tea garden intercropping with soybeans. The tillage depth control system comprises a body posture sensor, a control unit, and a hybrid stepper motor, integrating sensor data to drive the motor and achieve precise depth control. Real-time displacement sensor signals are compared with target values, enabling closed-loop control of the rotary tiller.
View Article and Find Full Text PDFSensors (Basel)
January 2025
Huawei Technologies Co., Ltd., Chengdu 610000, China.
Metasurface-based imaging is attractive due to its low hardware costs and system complexity. However, most of the current metasurface-based imaging systems require stochastic wavefront modulation, complex computational post-processing, and are restricted to 2D imaging. To overcome these limitations, we propose a scanning virtual aperture imaging system.
View Article and Find Full Text PDFSensors (Basel)
January 2025
Faculty of Science and Technology, Keio University, Yokohama 223-8522, Japan.
Person identification is a critical task in applications such as security and surveillance, requiring reliable systems that perform robustly under diverse conditions. This study evaluates the Vision Transformer (ViT) and ResNet34 models across three modalities-RGB, thermal, and depth-using datasets collected with infrared array sensors and LiDAR sensors in controlled scenarios and varying resolutions (16 × 12 to 640 × 480) to explore their effectiveness in person identification. Preprocessing techniques, including YOLO-based cropping, were employed to improve subject isolation.
View Article and Find Full Text PDFSensors (Basel)
January 2025
Meteorology and Fluid Science Division, Central Research Institute of Electric Power Industry, 1646 Abiko, Abiko-shi 270-1194, Chiba, Japan.
The electrical resistance (ER) method is widely used for atmospheric corrosion measurements and can be used to measure the corrosion rate accurately. However, severe errors occur in environments with temperature fluctuations, such as areas exposed to solar radiation, preventing accurate temporal corrosion rate measurement. To decrease the error, we developed an improved sensor composed of a reference metal film and an overlaid sensor metal film to cancel temperature differences between them.
View Article and Find Full Text PDFSensors (Basel)
January 2025
School of Computer Science, Hubei University of Technology, Wuhan 430068, China.
Large visual language models like Contrastive Language-Image Pre-training (CLIP), despite their excellent performance, are highly vulnerable to the influence of adversarial examples. This work investigates the accuracy and robustness of visual language models (VLMs) from a novel multi-modal perspective. We propose a multi-modal fine-tuning method called Multi-modal Depth Adversarial Prompt Tuning (MDAPT), which guides the generation of visual prompts through text prompts to improve the accuracy and performance of visual language models.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!