Pedestrian detection is widely used in real-time surveillance, urban traffic, and other fields. As a crucial direction in pedestrian detection, dense pedestrian detection still faces many unresolved challenges. Existing methods suffer from low detection accuracy, high miss rates, large model parameters, and poor robustness. In this paper, to address these issues, we propose a lightweight dense pedestrian detection model with finer-grained feature information interaction called MSCD-YOLO, which can achieve high accuracy, high performance and robustness with only a small number of parameters. In our model, the light-weight backbone network MobileViT is used to reduce the number of parameters while efficiently extracting both local and global features; the SCNeck neck network is designed to fuse the extracted features without losing information; and the DEHead detection head is utilized for multi-scale feature fusion to detect the targets. To demonstrate the effectiveness of our model, we conducted tests on the highly challenging dense pedestrian detection datasets Crowdhuman and Widerperson. Compared to the baseline model YOLOv8n, MSCD-YOLO achieved a 4.6% and 1.8% improvement in mAP@0.5, and a 5.3% and 2.6% improvement in mAP@0.5:0.95 on the Crowdhuman and Widerperson datasets, respectively. The experimental results show that under the same experimental conditions, MSCD-YOLO significantly outperforms the original model in terms of detection accuracy, efficiency, and model complexity.

Download full-text PDF

Source
http://dx.doi.org/10.3390/s25020438DOI Listing

Publication Analysis

Top Keywords

pedestrian detection
24
dense pedestrian
16
detection
9
lightweight dense
8
model
8
detection model
8
model finer-grained
8
finer-grained feature
8
feature interaction
8
detection accuracy
8

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!