Tony Felefly, Ziad Francis, Camille Roukoz, Georges Fares, Samir Achkar, Sandrine Yazbeck, Antoine Nasr, Manal Kordahi, Fares Azoury, Dolly Nehme Nasr, Elie Nasr, Georges Noël
{"title":"A 3D Convolutional Neural Network Based on Non-enhanced Brain CT to Identify Patients with Brain Metastases.","authors":"Tony Felefly, Ziad Francis, Camille Roukoz, Georges Fares, Samir Achkar, Sandrine Yazbeck, Antoine Nasr, Manal Kordahi, Fares Azoury, Dolly Nehme Nasr, Elie Nasr, Georges Noël","doi":"10.1007/s10278-024-01240-5","DOIUrl":"10.1007/s10278-024-01240-5","url":null,"abstract":"<p><p>Dedicated brain imaging for cancer patients is seldom recommended in the absence of symptoms. There is increasing availability of non-enhanced CT (NE-CT) of the brain, mainly owing to a wider utilization of Positron Emission Tomography-CT (PET-CT) in cancer staging. Brain metastases (BM) are often hard to diagnose on NE-CT. This work aims to develop a 3D Convolutional Neural Network (3D-CNN) based on brain NE-CT to distinguish patients with and without BM. We retrospectively included NE-CT scans for 100 patients with single or multiple BM and 100 patients without brain imaging abnormalities. Patients whose largest lesion was < 5 mm were excluded. The largest tumor was manually segmented on a matched contrast-enhanced T1 weighted Magnetic Resonance Imaging (MRI), and shape radiomics were extracted to determine the size and volume of the lesion. The brain was automatically segmented, and masked images were normalized and resampled. The dataset was split into training (70%) and validation (30%) sets. Multiple versions of a 3D-CNN were developed, and the best model was selected based on accuracy (ACC) on the validation set. The median largest tumor Maximum-3D-Diameter was 2.29 cm, and its median volume was 2.81 cc. Solitary BM were found in 27% of the patients, while 49% had > 5 BMs. The best model consisted of 4 convolutional layers with 3D average pooling layers, dropout layers of 50%, and a sigmoid activation function. Mean validation ACC was 0.983 (SD: 0.020) and mean area under receiver-operating characteristic curve was 0.983 (SD: 0.023). Sensitivity was 0.983 (SD: 0.020). We developed an accurate 3D-CNN based on brain NE-CT to differentiate between patients with and without BM. The model merits further external validation.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"858-864"},"PeriodicalIF":0.0,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11950574/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142074924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"DECNet: Left Atrial Pulmonary Vein Class Imbalance Classification Network.","authors":"GuoDong Zhang, WenWen Gu, TingYu Liang, YanLin Li, Wei Guo, ZhaoXuan Gong, RongHui Ju","doi":"10.1007/s10278-024-01221-8","DOIUrl":"10.1007/s10278-024-01221-8","url":null,"abstract":"<p><p>In clinical practice, the anatomical classification of pulmonary veins plays a crucial role in the preoperative assessment of atrial fibrillation radiofrequency ablation surgery. Accurate classification of pulmonary vein anatomy assists physicians in selecting appropriate mapping electrodes and avoids causing pulmonary arterial hypertension. Due to the diverse and subtly different anatomical classifications of pulmonary veins, as well as the imbalance in data distribution, deep learning models often exhibit poor expression capability in extracting deep features, leading to misjudgments and affecting classification accuracy. Therefore, in order to solve the problem of unbalanced classification of left atrial pulmonary veins, this paper proposes a network integrating multi-scale feature-enhanced attention and dual-feature extraction classifiers, called DECNet. The multi-scale feature-enhanced attention utilizes multi-scale information to guide the reinforcement of deep features, generating channel weights and spatial weights to enhance the expression capability of deep features. The dual-feature extraction classifier assigns a fixed number of channels to each category, equally evaluating all categories, thus alleviating the learning bias and overfitting caused by data imbalance. By combining the two, the expression capability of deep features is strengthened, achieving accurate classification of left atrial pulmonary vein morphology and providing support for subsequent clinical treatment. The proposed method is evaluated on datasets provided by the People's Hospital of Liaoning Province and the publicly available DermaMNIST dataset, achieving average accuracies of 78.81% and 83.44%, respectively, demonstrating the effectiveness of the proposed approach.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"819-837"},"PeriodicalIF":0.0,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11950506/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142010198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Deep Learning-Based Model for Non-invasive Hemoglobin Estimation via Body Parts Images: A Retrospective Analysis and a Prospective Emergency Department Study.","authors":"En-Ting Lin, Shao-Chi Lu, An-Sheng Liu, Chia-Hsin Ko, Chien-Hua Huang, Chu-Lin Tsai, Li-Chen Fu","doi":"10.1007/s10278-024-01209-4","DOIUrl":"10.1007/s10278-024-01209-4","url":null,"abstract":"<p><p>Anemia is a significant global health issue, affecting over a billion people worldwide, according to the World Health Organization. Generally, the gold standard for diagnosing anemia relies on laboratory measurements of hemoglobin. To meet the need in clinical practice, physicians often rely on visual examination of specific areas, such as conjunctiva, to assess pallor. However, this method is subjective and relies on the physician's experience. Therefore, we proposed a deep learning prediction model based on three input images from different body parts, namely, conjunctiva, palm, and fingernail. By incorporating additional body part labels and employing a fusion attention mechanism, the model learns and enhances the salient features of each body part during training, enabling it to produce reliable results. Additionally, we employ a dual loss function that allows the regression model to benefit from well-established classification methods, thereby achieving stable handling of minority samples. We used a retrospective data set (EYES-DEFY-ANEMIA) to develop this model called Body-Part-Anemia Network (BPANet). The BPANet showed excellent performance in detecting anemia, with accuracy of 0.849 and an F1-score of 0.828. Our multi-body-part model has been validated on a prospectively collected data set of 101 patients in National Taiwan University Hospital. The prediction accuracy as well as F1-score can achieve as high as 0.716 and 0.788, respectively. To sum up, we have developed and validated a novel non-invasive hemoglobin prediction model based on image input from multiple body parts, with the potential of real-time use at home and in clinical settings.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"775-792"},"PeriodicalIF":0.0,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11950610/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142006261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Application of Wiener Filter Based on Improved BB Gradient Descent in Iris Image Restoration.","authors":"Chuandong Qin, Yiqing Zhang","doi":"10.1007/s10278-024-01238-z","DOIUrl":"10.1007/s10278-024-01238-z","url":null,"abstract":"<p><p>Iris recognition, renowned for its exceptional precision, has been extensively utilized across diverse industries. However, the presence of noise and blur frequently compromises the quality of iris images, thereby adversely affecting recognition accuracy. In this research, we have refined the traditional Wiener filter image restoration technique by integrating it with a gradient descent strategy, specifically employing the Barzilai-Borwein (BB) step size selection. This innovative approach is designed to enhance both the precision and resilience of iris recognition systems. The BB gradient method is adept at optimizing the parameters of the Wiener filter by introducing simulated blurring and noise conditions to the iris images. Through this process, it is capable of restoring images that have been degraded by blur and noise, leading to a significant improvement in the clarity of the restored images and, consequently, a notable elevation in recognition performance. The results of our experiments have demonstrated that this advanced method surpasses conventional filtering techniques in terms of both subjective visual quality assessments and objective peak signal-to-noise ratio (PSNR) evaluations.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"1165-1183"},"PeriodicalIF":0.0,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11950590/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142134996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Mitchell Goldburgh, Michael LaChance, Julia Komissarchik, Julia Patriarche, Joe Chapa, Oliver Chen, Priya Deshpande, Matthew Geeslin, Nina Kottler, Jennifer Sommer, Marcus Ayers, Vedrana Vujic
{"title":"2023 Industry Perceptions Survey on AI Adoption and Return on Investment.","authors":"Mitchell Goldburgh, Michael LaChance, Julia Komissarchik, Julia Patriarche, Joe Chapa, Oliver Chen, Priya Deshpande, Matthew Geeslin, Nina Kottler, Jennifer Sommer, Marcus Ayers, Vedrana Vujic","doi":"10.1007/s10278-024-01147-1","DOIUrl":"10.1007/s10278-024-01147-1","url":null,"abstract":"<p><p>This SIIM-sponsored 2023 report highlights an industry view on artificial intelligence adoption barriers and success related to diagnostic imaging, life sciences, and contrasts. In general, our 2023 survey indicates that there has been progress in adopting AI across multiple uses, and there continues to be an optimistic forecast for the impact on workflow and clinical outcomes. This report, as in prior years, should be seen as a snapshot of the use of AI in imaging. Compared to our 2021 survey, the 2023 respondents expressed wider AI adoption but felt this was behind the potential. Specifically, the adoption has increased as sources of return on investment with AI in radiology are better understood as documented by vendor/client use case studies. Generally, the discussions of AI solutions centered on workflow triage, visualization, detection, and characterization. Generative AI was also mentioned for improving productivity in reporting. As payor reimbursement remains elusive, the ROI discussions expanded to look at other factors, including increased hospital procedures and admissions, enhanced radiologist productivity for practices, and improved patient outcomes for integrated health networks. When looking at the longer-term horizon for AI adoption, respondents frequently mentioned that the opportunity for AI to achieve greater adoption with more complex AI and a more manageable/visible ROI is outside the USA. Respondents focused on the barriers to trust in AI and the FDA processes.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"663-670"},"PeriodicalIF":0.0,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11950608/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142010196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Mohammadreza Zandehshahvar, Marly van Assen, Eun Kim, Yashar Kiarashi, Vikranth Keerthipati, Giovanni Tessarin, Emanuele Muscogiuri, Arthur E Stillman, Peter Filev, Amir H Davarpanah, Eugene A Berkowitz, Stefan Tigges, Scott J Lee, Brianna L Vey, Carlo De Cecco, Ali Adibi
{"title":"Confidence-Aware Severity Assessment of Lung Disease from Chest X-Rays Using Deep Neural Network on a Multi-Reader Dataset.","authors":"Mohammadreza Zandehshahvar, Marly van Assen, Eun Kim, Yashar Kiarashi, Vikranth Keerthipati, Giovanni Tessarin, Emanuele Muscogiuri, Arthur E Stillman, Peter Filev, Amir H Davarpanah, Eugene A Berkowitz, Stefan Tigges, Scott J Lee, Brianna L Vey, Carlo De Cecco, Ali Adibi","doi":"10.1007/s10278-024-01151-5","DOIUrl":"10.1007/s10278-024-01151-5","url":null,"abstract":"<p><p>In this study, we present a method based on Monte Carlo Dropout (MCD) as Bayesian neural network (BNN) approximation for confidence-aware severity classification of lung diseases in COVID-19 patients using chest X-rays (CXRs). Trained and tested on 1208 CXRs from Hospital 1 in the USA, the model categorizes severity into four levels (i.e., normal, mild, moderate, and severe) based on lung consolidation and opacity. Severity labels, determined by the median consensus of five radiologists, serve as the reference standard. The model's performance is internally validated against evaluations from an additional radiologist and two residents that were excluded from the median. The performance of the model is further evaluated on additional internal and external datasets comprising 2200 CXRs from the same hospital and 1300 CXRs from Hospital 2 in South Korea. The model achieves an average area under the curve (AUC) of 0.94 ± 0.01 across all classes in the primary dataset, surpassing human readers in each severity class and achieves a higher Kendall correlation coefficient (KCC) of 0.80 ± 0.03. The performance of the model is consistent across varied datasets, highlighting its generalization. A key aspect of the model is its predictive uncertainty (PU), which is inversely related to the level of agreement among radiologists, particularly in mild and moderate cases. The study concludes that the model outperforms human readers in severity assessment and maintains consistent accuracy across diverse datasets. Its ability to provide confidence measures in predictions is pivotal for potential clinical use, underscoring the BNN's role in enhancing diagnostic precision in lung disease analysis through CXR.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"793-803"},"PeriodicalIF":0.0,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11950521/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142010197","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Dual-Tree Complex Wavelet Pooling and Attention-Based Modified U-Net Architecture for Automated Breast Thermogram Segmentation and Classification.","authors":"Lalit Garia, Hariharan Muthusamy","doi":"10.1007/s10278-024-01239-y","DOIUrl":"10.1007/s10278-024-01239-y","url":null,"abstract":"<p><p>Thermography is a non-invasive and non-contact method for detecting cancer in its initial stages by examining the temperature variation between both breasts. Preprocessing methods such as resizing, ROI (region of interest) segmentation, and augmentation are frequently used to enhance the accuracy of breast thermogram analysis. In this study, a modified U-Net architecture (DTCWAU-Net) that uses dual-tree complex wavelet transform (DTCWT) and attention gate for breast thermal image segmentation for frontal and lateral view thermograms, aiming to outline ROI for potential tumor detection, was proposed. The proposed approach achieved an average Dice coefficient of 93.03% and a sensitivity of 94.82%, showcasing its potential for accurate breast thermogram segmentation. Classification of breast thermograms into healthy or cancerous categories was carried out by extracting texture- and histogram-based features and deep features from segmented thermograms. Feature selection was performed using Neighborhood Component Analysis (NCA), followed by the application of machine learning classifiers. When compared to other state-of-the-art approaches for detecting breast cancer using a thermogram, the proposed methodology showed a higher accuracy of 99.90% for VGG16 deep features with NCA and Random Forest classifier. Simulation results expound that the proposed method can be used in breast cancer screening, facilitating early detection, and enhancing treatment outcomes.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"887-901"},"PeriodicalIF":0.0,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11950499/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142127894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Interactive Multi-scale Fusion: Advancing Brain Tumor Detection Through Trans-IMSM Model.","authors":"Vasanthi Durairaj, Palani Uthirapathy","doi":"10.1007/s10278-024-01222-7","DOIUrl":"10.1007/s10278-024-01222-7","url":null,"abstract":"<p><p>Multi-modal medical image (MI) fusion assists in generating collaboration images collecting complement features through the distinct images of several conditions. The images help physicians to diagnose disease accurately. Hence, this research proposes a novel multi-modal MI fusion modal named guided filter-based interactive multi-scale and multi-modal transformer (Trans-IMSM) fusion approach to develop high-quality computed tomography-magnetic resonance imaging (CT-MRI) fused images for brain tumor detection. This research utilizes the CT and MRI brain scan dataset to gather the input CT and MRI images. At first, the data preprocessing is carried out to preprocess these input images to improve the image quality and generalization ability for further analysis. Then, these preprocessed CT and MRI are decomposed into detail and base components utilizing the guided filter-based MI decomposition approach. This approach involves two phases: such as acquiring the image guidance and decomposing the images utilizing the guided filter. A canny operator is employed to acquire the image guidance comprising robust edge for CT and MRI images, and the guided filter is applied to decompose the guidance and preprocessed images. Then, by applying the Trans-IMSM model, fuse the detail components, while a weighting approach is used for the base components. The fused detail and base components are subsequently processed through a gated fusion and reconstruction network, and the final fused images for brain tumor detection are generated. Extensive tests are carried out to compute the Trans-IMSM method's efficacy. The evaluation results demonstrated the robustness and effectiveness, achieving an accuracy of 98.64% and an SSIM of 0.94.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"757-774"},"PeriodicalIF":0.0,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11950544/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141989828","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A Deep-Learning-Enabled Electrocardiogram and Chest X-Ray for Detecting Pulmonary Arterial Hypertension.","authors":"Pang-Yen Liu, Shi-Chue Hsing, Dung-Jang Tsai, Chin Lin, Chin-Sheng Lin, Chih-Hung Wang, Wen-Hui Fang","doi":"10.1007/s10278-024-01225-4","DOIUrl":"10.1007/s10278-024-01225-4","url":null,"abstract":"<p><p>The diagnosis and treatment of pulmonary hypertension have changed dramatically through the re-defined diagnostic criteria and advanced drug development in the past decade. The application of Artificial Intelligence for the detection of elevated pulmonary arterial pressure (ePAP) was reported recently. Artificial Intelligence (AI) has demonstrated the capability to identify ePAP and its association with hospitalization due to heart failure when analyzing chest X-rays (CXR). An AI model based on electrocardiograms (ECG) has shown promise in not only detecting ePAP but also in predicting future risks related to cardiovascular mortality. We aimed to develop an AI model integrating ECG and CXR to detect ePAP and evaluate their performance. We developed a deep-learning model (DLM) using paired ECG and CXR to detect ePAP (systolic pulmonary artery pressure > 50 mmHg in transthoracic echocardiography). This model was further validated in a community hospital. Additionally, our DLM was evaluated for its ability to predict future occurrences of left ventricular dysfunction (LVD, ejection fraction < 35%) and cardiovascular mortality. The AUCs for detecting ePAP were as follows: 0.8261 with ECG (sensitivity 76.6%, specificity 74.5%), 0.8525 with CXR (sensitivity 82.8%, specificity 72.7%), and 0.8644 with a combination of both (sensitivity 78.6%, specificity 79.2%) in the internal dataset. In the external validation dataset, the AUCs for ePAP detection were 0.8348 with ECG, 0.8605 with CXR, and 0.8734 with the combination. Furthermore, using the combination of ECGs and CXR, the negative predictive value (NPV) was 98% in the internal dataset and 98.1% in the external dataset. Patients with ePAP detected by the DLM using combination had a higher risk of new-onset LVD with a hazard ratio (HR) of 4.51 (95% CI: 3.54-5.76) in the internal dataset and cardiovascular mortality with a HR of 6.08 (95% CI: 4.66-7.95). Similar results were seen in the external validation dataset. The DLM, integrating ECG and CXR, effectively detected ePAP with a strong NPV and forecasted future risks of developing LVD and cardiovascular mortality. This model has the potential to expedite the early identification of pulmonary hypertension in patients, prompting further evaluation through echocardiography and, when necessary, right heart catheterization (RHC), potentially resulting in enhanced cardiovascular outcomes.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"747-756"},"PeriodicalIF":0.0,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11950589/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141972525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"From Revisions to Insights: Converting Radiology Report Revisions into Actionable Educational Feedback Using Generative AI Models.","authors":"Shawn Lyo, Suyash Mohan, Alvand Hassankhani, Abass Noor, Farouk Dako, Tessa Cook","doi":"10.1007/s10278-024-01233-4","DOIUrl":"10.1007/s10278-024-01233-4","url":null,"abstract":"<p><p>Expert feedback on trainees' preliminary reports is crucial for radiologic training, but real-time feedback can be challenging due to non-contemporaneous, remote reading and increasing imaging volumes. Trainee report revisions contain valuable educational feedback, but synthesizing data from raw revisions is challenging. Generative AI models can potentially analyze these revisions and provide structured, actionable feedback. This study used the OpenAI GPT-4 Turbo API to analyze paired synthesized and open-source analogs of preliminary and finalized reports, identify discrepancies, categorize their severity and type, and suggest review topics. Expert radiologists reviewed the output by grading discrepancies, evaluating the severity and category accuracy, and suggested review topic relevance. The reproducibility of discrepancy detection and maximal discrepancy severity was also examined. The model exhibited high sensitivity, detecting significantly more discrepancies than radiologists (W = 19.0, p < 0.001) with a strong positive correlation (r = 0.778, p < 0.001). Interrater reliability for severity and type were fair (Fleiss' kappa = 0.346 and 0.340, respectively; weighted kappa = 0.622 for severity). The LLM achieved a weighted F1 score of 0.66 for severity and 0.64 for type. Generated teaching points were considered relevant in ~ 85% of cases, and relevance correlated with the maximal discrepancy severity (Spearman ρ = 0.76, p < 0.001). The reproducibility was moderate to good (ICC (2,1) = 0.690) for the number of discrepancies and substantial for maximal discrepancy severity (Fleiss' kappa = 0.718; weighted kappa = 0.94). Generative AI models can effectively identify discrepancies in report revisions and generate relevant educational feedback, offering promise for enhancing radiology training.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"1265-1279"},"PeriodicalIF":0.0,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11950553/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142006262","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}