Recognizing and monitoring infectious sources of schistosomiasis by developing deep learning models with high-resolution remote sensing images.

Infect Dis Poverty

National Institute of Parasitic Diseases, Chinese Center for Disease Control and Prevention (Chinese Center for Tropical Diseases Research); NHC Key Laboratory of Parasite and Vector Biology; WHO Collaborating Centre for Tropical Diseases, National Center for International Research On Tropical Diseases, Shanghai, 200025, China.

Published: February 2023

AI Article Synopsis

  • China is working towards eliminating schistosomiasis, but faces challenges in managing infection sources and controlling snails, leading to a study that uses deep learning and remote sensing to track livestock, specifically bovine, which spread the disease.
  • Researchers developed two deep learning models, ENVINet5 and Mask R-CNN, using high-resolution images to identify and monitor bovine populations, achieving promising precision and recall metrics of up to 87.3% and 85.2%, respectively.
  • When applied to real-world scenarios, the Mask R-CNN model performed better with a recognition rate of 90.5% for bovine in schistosomiasis-prone areas, demonstrating its potential for effective disease

Article Abstract

Background: China is progressing towards the goal of schistosomiasis elimination, but there are still some problems, such as difficult management of infection source and snail control. This study aimed to develop deep learning models with high-resolution remote sensing images for recognizing and monitoring livestock bovine, which is an intermediate source of Schistosoma japonicum infection, and to evaluate the effectiveness of the models for real-world application.

Methods: The dataset of livestock bovine's spatial distribution was collected from the Chinese National Platform for Common Geospatial Information Services. The high-resolution remote sensing images were further divided into training data, test data, and validation data for model development. Two recognition models based on deep learning methods (ENVINet5 and Mask R-CNN) were developed with reference to the training datasets. The performance of the developed models was evaluated by the performance metrics of precision, recall, and F1-score.

Results: A total of 50 typical image areas were selected, 1125 bovine objectives were labeled by the ENVINet5 model and 1277 bovine objectives were labeled by the Mask R-CNN model. For the ENVINet5 model, a total of 1598 records of bovine distribution were recognized. The model precision and recall were 81.9% and 80.2%, respectively. The F1 score was 0.81. For the Mask R-CNN mode, 1679 records of bovine objectives were identified. The model precision and recall were 87.3% and 85.2%, respectively. The F1 score was 0.87. When applying the developed models to real-world schistosomiasis-endemic regions, there were 63 bovine objectives in the original image, 53 records were extracted using the ENVINet5 model, and 57 records were extracted using the Mask R-CNN model. The successful recognition ratios were 84.1% and 90.5% for the respectively developed models.

Conclusion: The ENVINet5 model is very feasible when the bovine distribution is low in structure with few samples. The Mask R-CNN model has a good framework design and runs highly efficiently. The livestock recognition models developed using deep learning methods with high-resolution remote sensing images accurately recognize the spatial distribution of livestock, which could enable precise control of schistosomiasis.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC9903608PMC
http://dx.doi.org/10.1186/s40249-023-01060-9DOI Listing

Publication Analysis

Top Keywords

mask r-cnn
20
deep learning
16
high-resolution remote
16
remote sensing
16
sensing images
16
bovine objectives
16
envinet5 model
16
precision recall
12
r-cnn model
12
model
10

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!