A PHP Error was encountered

Severity: Warning

Message: file_get_contents(https://...@gmail.com&api_key=61f08fa0b96a73de8c900d749fcb997acc09&a=1): Failed to open stream: HTTP request failed! HTTP/1.1 429 Too Many Requests

Filename: helpers/my_audit_helper.php

Line Number: 197

Backtrace:

File: /var/www/html/application/helpers/my_audit_helper.php
Line: 197
Function: file_get_contents

File: /var/www/html/application/helpers/my_audit_helper.php
Line: 271
Function: simplexml_load_file_from_url

File: /var/www/html/application/helpers/my_audit_helper.php
Line: 3145
Function: getPubMedXML

File: /var/www/html/application/controllers/Detail.php
Line: 575
Function: pubMedSearch_Global

File: /var/www/html/application/controllers/Detail.php
Line: 489
Function: pubMedGetRelatedKeyword

File: /var/www/html/index.php
Line: 316
Function: require_once

Analysis of ILD-RADS reproducibility and validity: Can different degrees of experience affect the interpretation? | LitMetric

Purpose: To evaluate the reliability, validity and applicability of ILD-RADS among readers with different levels of experience.

Methods: This prospective tri-center study included 159 patients with clinically diagnosed ILD who underwent high-resolution CT (HRCT). Two experienced thoracic radiologists, two general radiologists, and one pulmonologist independently evaluated the HRCT images blinded to the patient's clinical data and assigned ILD-RADS category for each patient. The Fleiss kappa test was employed to estimate the inter-reader agreement among all readers. Cohen's kappa test was applied to measure the pairwise inter-reader agreement. The multi-disciplinary team discussion (MDD) was used as a reference test to estimate the validity of ILD-RADS for diagnosing idiopathic pulmonary fibrosis (IPF). A 5-point Likert short survey was accomplished by the pulmonologists about the applicability of ILD-RADS in clinical practice.

Results: The current study included 124 non-IPF and 35 IPF cases. Based on the radiologists, the ILD-RADS showed moderate inter-reader agreement (K = 0.515, P < 0.001) while being fair after the inclusion of the pulmonologist's input (K = 0.333, P < 0.001). The agreement was substantial among thoracic radiologists (K = 0.716, p < 0.001) and moderate among general radiologists (K = 0.461, p < 0.001). ILD-RADS ≤ 2 was the optimal cut-off for predicting IPF, with an accuracy ranging from 62.84 % to 80.54 %. Seventy-five percent of pulmonologists rated ILD-RADS as highly applicable in practice.

Conclusions: ILD-RADS is reliable and valid among radiologists but requires further refinement to enhance consistency and applicability in diverse clinical settings. Moreover, pulmonologists support its use in clinical practice.

Download full-text PDF

Source
http://dx.doi.org/10.1016/j.ejrad.2025.111961DOI Listing

Publication Analysis

Top Keywords

inter-reader agreement
12
applicability ild-rads
8
study included
8
kappa test
8
ild-rads
5
analysis ild-rads
4
ild-rads reproducibility
4
reproducibility validity
4
validity degrees
4
degrees experience
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!