A Two-Phase Cross-Modality Fusion Network for Robust 3D Object Detection.

Sensors (Basel)

School of Automotive Engineering, Wuhan University of Technology, Wuhan 430070, China.

Published: October 2020

A two-phase cross-modality fusion detector is proposed in this study for robust and high-precision 3D object detection with RGB images and LiDAR point clouds. First, a two-stream fusion network is built into the framework of Faster RCNN to perform accurate and robust 2D detection. The visible stream takes the RGB images as inputs, while the intensity stream is fed with the intensity maps which are generated by projecting the reflection intensity of point clouds to the front view. A multi-layer feature-level fusion scheme is designed to merge multi-modal features across multiple layers in order to enhance the expressiveness and robustness of the produced features upon which region proposals are generated. Second, a decision-level fusion is implemented by projecting 2D proposals to the space of the point cloud to generate 3D frustums, on the basis of which the second-phase 3D detector is built to accomplish instance segmentation and 3D-box regression on the filtered point cloud. The results on the KITTI benchmark show that features extracted from RGB images and intensity maps complement each other, and our proposed detector achieves state-of-the-art performance on 3D object detection with a substantially lower running time as compared to available competitors.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC7660652PMC
http://dx.doi.org/10.3390/s20216043DOI Listing

Publication Analysis

Top Keywords

object detection
12
rgb images
12
two-phase cross-modality
8
cross-modality fusion
8
fusion network
8
point clouds
8
intensity maps
8
point cloud
8
fusion
5
network robust
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!