Machine-learned models provided substantial gains in model sensitivity with slight loss of specificity.
A natural language processing (NLP) system helped identify lumbar spine findings, providing substantial gains in model sensitivity, according to a study published in the journal Academic Radiology.
Researchers from several states sought to evaluate an NLP system built with open-source tools for identification of lumbar spine imaging findings related to low back pain on magnetic resonance and x-ray radiology reports from four health systems.
The researchers used a limited data set sampled from lumbar spine imaging reports of a prospectively assembled cohort of adults. A total of 871 reports were randomly selected from 178,333 available reports; 413 were x-rays and 458 were MR reports.
Related article: Are You Tired of Repeating Yourself?
Using standardized criteria, four spine experts annotated the presence of 26 findings, where 71 reports were annotated by all four experts and 800 were each annotated by two experts. The researchers calculated inter-rater agreement and finding prevalence from annotated data. The annotated data was randomly split into development (80%) and testing (20%) sets. The researchers developed an NLP system from both rule-based and machine-learned models. The system was validated using accuracy metrics such as sensitivity, specificity, and area under the receiver operating characteristic curve (AUC).
The results showed the multirater annotated dataset achieved inter-rater agreement of Cohen's kappa > 0.60 (substantial agreement) for 25 of 26 findings, with finding prevalence ranging from 3% to 89%. In the testing sample, rule-based and machine-learned predictions both had comparable average specificity (0.97 and 0.95, respectively). The machine-learned approach had a higher average sensitivity (0.94, compared to 0.83 for rules-based), and a higher overall AUC (0.98, compared to 0.90 for rules-based).
The researchers concluded that their NLP system performed well in identifying the 26 lumbar spine findings, as benchmarked by reference-standard annotation by medical experts. Machine-learned models provided substantial gains in model sensitivity with slight loss of specificity, and overall higher AUC.
New Study Examines Short-Term Consistency of Large Language Models in Radiology
November 22nd 2024While GPT-4 demonstrated higher overall accuracy than other large language models in answering ACR Diagnostic in Training Exam multiple-choice questions, researchers noted an eight percent decrease in GPT-4’s accuracy rate from the first month to the third month of the study.
The Reading Room: Artificial Intelligence: What RSNA 2020 Offered, and What 2021 Could Bring
December 5th 2020Nina Kottler, M.D., chief medical officer of AI at Radiology Partners, discusses, during RSNA 2020, what new developments the annual meeting provided about these technologies, sessions to access, and what to expect in the coming year.