{"title":"Multimodal Deep Learning for Grading Carpal Tunnel Syndrome: A Multicenter Study in China.","authors":"Xiaochen Shi, Tianxiang Yu, Yu Yuan, Dan Wang, Jinhua Cui, Ling Bai, Fang Zheng, Xiaobin Dai, Zhuhuang Zhou","doi":"10.1016/j.acra.2025.02.043","DOIUrl":null,"url":null,"abstract":"<p><strong>Rationale and objectives: </strong>Ultrasound (US)-based deep learning (DL) models for grading the severity of carpal tunnel syndrome (CTS) are scarce. We aimed to advance CTS grading by developing a joint-DL model integrating clinical information and multimodal US features.</p><p><strong>Materials and methods: </strong>A retrospective dataset of CTS patients from three hospitals was randomly divided into the training (n=680) and internal validation (n=173) sets. An external validation set was prospectively recruited from another hospital (n=174). To further test the model's generalizability, cross-vendor testing was conducted at three additional hospitals utilizing different US systems in the external validation set 2 (n=224). An US-based model was developed to grade CTS severity utilizing multimodal sonographic features, including cross-sectional area [CSA], echogenicity, longitudinal nerve appearance, and intraneural vascularity. A joint-DL model (CTSGrader) was constructed integrating sonographic features and clinical information. Diagnostic performance of both models was verified based on electrophysiological results. In the validation sets, the better-performing model was compared to two junior and two senior radiologists. Additionally, the radiologists' diagnostic performance with artificial intelligence (AI) assistance was evaluated in external validation sets.</p><p><strong>Results: </strong>CTSGrader achieved areas under the curve (AUCs) of 0.951, 0.910, and 0.897 in the validation sets. The accuracies of CTSGrader were 0.849, 0.833, and 0.827, which were higher than those of US-based model (all p<.05). It outperformed two junior and one senior radiologists (all p<.05) and was equivalent to 1 senior radiologist (all p>.05). With its assistance, the accuracies of two junior and one senior radiologists were improved (all p<.05).</p><p><strong>Conclusion: </strong>The joint-DL model (CTSGrader) developed in our study outperformed the single-modality model. The AI-aided strategy suggested its potential to support clinical decision-making for grading CTS severity.</p>","PeriodicalId":50928,"journal":{"name":"Academic Radiology","volume":" ","pages":""},"PeriodicalIF":3.8000,"publicationDate":"2025-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Academic Radiology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1016/j.acra.2025.02.043","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0
Abstract
Rationale and objectives: Ultrasound (US)-based deep learning (DL) models for grading the severity of carpal tunnel syndrome (CTS) are scarce. We aimed to advance CTS grading by developing a joint-DL model integrating clinical information and multimodal US features.
Materials and methods: A retrospective dataset of CTS patients from three hospitals was randomly divided into the training (n=680) and internal validation (n=173) sets. An external validation set was prospectively recruited from another hospital (n=174). To further test the model's generalizability, cross-vendor testing was conducted at three additional hospitals utilizing different US systems in the external validation set 2 (n=224). An US-based model was developed to grade CTS severity utilizing multimodal sonographic features, including cross-sectional area [CSA], echogenicity, longitudinal nerve appearance, and intraneural vascularity. A joint-DL model (CTSGrader) was constructed integrating sonographic features and clinical information. Diagnostic performance of both models was verified based on electrophysiological results. In the validation sets, the better-performing model was compared to two junior and two senior radiologists. Additionally, the radiologists' diagnostic performance with artificial intelligence (AI) assistance was evaluated in external validation sets.
Results: CTSGrader achieved areas under the curve (AUCs) of 0.951, 0.910, and 0.897 in the validation sets. The accuracies of CTSGrader were 0.849, 0.833, and 0.827, which were higher than those of US-based model (all p<.05). It outperformed two junior and one senior radiologists (all p<.05) and was equivalent to 1 senior radiologist (all p>.05). With its assistance, the accuracies of two junior and one senior radiologists were improved (all p<.05).
Conclusion: The joint-DL model (CTSGrader) developed in our study outperformed the single-modality model. The AI-aided strategy suggested its potential to support clinical decision-making for grading CTS severity.
期刊介绍:
Academic Radiology publishes original reports of clinical and laboratory investigations in diagnostic imaging, the diagnostic use of radioactive isotopes, computed tomography, positron emission tomography, magnetic resonance imaging, ultrasound, digital subtraction angiography, image-guided interventions and related techniques. It also includes brief technical reports describing original observations, techniques, and instrumental developments; state-of-the-art reports on clinical issues, new technology and other topics of current medical importance; meta-analyses; scientific studies and opinions on radiologic education; and letters to the Editor.