Rethinking Greulich and Pyle: A Deep Learning Approach to Pediatric Bone Age Assessment Using Pediatric Trauma Hand Radiographs.

Radiol Artif Intell

Department of Diagnostic Imaging, Rhode Island Hospital/Hasbro Children's Hospital, The Warren Alpert Medical School of Brown University, 593 Eddy St, Providence, RI 02903 (I.P., D.W.S., R.S.A.); Department of Diagnostic Imaging and Lifespan Biostatistics Core, Rhode Island Hospital, Providence, RI (G.L.B.); Department of Radiology, Columbia University Medical Center, New York, NY (S.M., C.R.); and Department of Emergency Medicine, University of Florida Shands Hospital, Gainesville, Fla (D.M.).

Published: July 2020

AI Article Synopsis

  • A deep learning model called TDL-BAAM was developed to assess bone age using 15,129 pediatric hand radiographs and was compared to both automated methods and manual assessments done by radiologists using the Greulich and Pyle (GP) method.
  • The TDL-BAAM model achieved a mean absolute error of 11.1 months, outperforming the GP-based model (12.9 months) and two pediatric radiologists (14.6 and 16.0 months).
  • The study concluded that the deep learning approach is comparable to traditional methods for bone age assessment, highlighting its potential for further development tailored to specific populations.

Article Abstract

Purpose: To develop a deep learning approach to bone age assessment based on a training set of developmentally normal pediatric hand radiographs and to compare this approach with automated and manual bone age assessment methods based on Greulich and Pyle (GP).

Methods: In this retrospective study, a convolutional neural network (trauma hand radiograph-trained deep learning bone age assessment method [TDL-BAAM]) was trained on 15 129 frontal view pediatric trauma hand radiographs obtained between December 14, 2009, and May 31, 2017, from Children's Hospital of New York, to predict chronological age. A total of 214 trauma hand radiographs from Hasbro Children's Hospital were used as an independent test set. The test set was rated by the TDL-BAAM model as well as a GP-based deep learning model (GPDL-BAAM) and two pediatric radiologists (radiologists 1 and 2) using the GP method. All ratings were compared with chronological age using mean absolute error (MAE), and standard concordance analyses were performed.

Results: The MAE of the TDL-BAAM model was 11.1 months, compared with 12.9 months for GPDL-BAAM ( = .0005), 14.6 months for radiologist 1 ( < .0001), and 16.0 for radiologist 2 ( < .0001). For TDL-BAAM, 95.3% of predictions were within 24 months of chronological age compared with 91.6% for GPDL-BAAM ( = .096), 86.0% for radiologist 1 ( < .0001), and 84.6% for radiologist 2 ( < .0001). Concordance was high between all methods and chronological age (intraclass coefficient > 0.93). Deep learning models demonstrated a systematic bias with a tendency to overpredict age for younger children versus radiologists who showed a consistent mean bias.

Conclusion: A deep learning model trained on pediatric trauma hand radiographs is on par with automated and manual GP-based methods for bone age assessment and provides a foundation for developing population-specific deep learning algorithms for bone age assessment in modern pediatric populations.© RSNA, 2020See also the commentary by Halabi in this issue.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC8082327PMC
http://dx.doi.org/10.1148/ryai.2020190198DOI Listing

Publication Analysis

Top Keywords

deep learning
28
bone age
24
age assessment
24
trauma hand
20
hand radiographs
20
chronological age
16
radiologist 0001
16
pediatric trauma
12
age
11
greulich pyle
8

Similar Publications

Purpose: Semantic segmentation and landmark detection are fundamental tasks of medical image processing, facilitating further analysis of anatomical objects. Although deep learning-based pixel-wise classification has set a new-state-of-the-art for segmentation, it falls short in landmark detection, a strength of shape-based approaches.

Methods: In this work, we propose a dense image-to-shape representation that enables the joint learning of landmarks and semantic segmentation by employing a fully convolutional architecture.

View Article and Find Full Text PDF

Currently, the World Health Organization (WHO) grade of meningiomas is determined based on the biopsy results. Therefore, accurate non-invasive preoperative grading could significantly improve treatment planning and patient outcomes. Considering recent advances in machine learning (ML) and deep learning (DL), this meta-analysis aimed to evaluate the performance of these models in predicting the WHO meningioma grade using imaging data.

View Article and Find Full Text PDF

Parkinson's disease (PD), a degenerative disorder of the central nervous system, is commonly diagnosed using functional medical imaging techniques such as single-photon emission computed tomography (SPECT). In this study, we utilized two SPECT data sets (n = 634 and n = 202) from different hospitals to develop a model capable of accurately predicting PD stages, a multiclass classification task. We used the entire three-dimensional (3D) brain images as input and experimented with various model architectures.

View Article and Find Full Text PDF

Accurate wound segmentation is crucial for the precise diagnosis and treatment of various skin conditions through image analysis. In this paper, we introduce a novel dual attention U-Net model designed for precise wound segmentation. Our proposed architecture integrates two widely used deep learning models, VGG16 and U-Net, incorporating dual attention mechanisms to focus on relevant regions within the wound area.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!