Alon Gorenshtein, Yana Weisblat, Mohamed Khateb, Gilad Kenan, Irina Tsirkin, Galina Fayn, Semion Geller, Shahar Shelly
{"title":"基于人工智能的肌电图报告:一项随机对照试验。","authors":"Alon Gorenshtein, Yana Weisblat, Mohamed Khateb, Gilad Kenan, Irina Tsirkin, Galina Fayn, Semion Geller, Shahar Shelly","doi":"10.1007/s00415-025-13261-3","DOIUrl":null,"url":null,"abstract":"<p><strong>Background and objectives: </strong>Accurate interpretation of electrodiagnostic (EDX) studies is essential for the diagnosis and management of neuromuscular disorders. Artificial intelligence (AI) based tools may improve consistency and quality of EDX reporting and reduce workload. The aim of this study is to evaluate the performance of an AI-assisted, multi-agent framework (INSPIRE) in comparison with standard physician interpretation in a randomized controlled trial (RCT).</p><p><strong>Methods: </strong>We prospectively enrolled 200 patients (out of 363 assessed for eligibility) referred for EDX. Patients were randomly assigned to either a control group (physician-only interpretation) or an intervention group (physician-AI interpretation). Three board-certified physicians, rotated across both arms. In the intervention group, an AI-generated preliminary report was combined with the physician's independent findings (human-AI integration). The primary outcome was EDX report quality using score we developed named AI-Generated EMG Report Score (AIGERS; score range, 0-1, with higher scores indicating more accurate or complete reports). Secondary outcomes included a physician-reported AI integration rating score (PAIR) and a compliance survey evaluating ease of AI adoption.</p><p><strong>Results: </strong>Of the 200 enrolled patients, 100 were allocated to AI-assisted interpretation and 100 to physician-only reporting. While AI-generated preliminary reports offered moderate consistency on the AIGERS metric, the integrated (physician-AI) approach did not significantly outperform physician-only methods. Despite some anecdotal advantages such as efficiency in suggesting standardized terminology quantitatively, the AIGERS scores for physician-AI integration was nearly the same as those in the physician-only arm and did not reach statistical significance (p > 0.05 for all comparisons). Physicians reported variable acceptance of AI suggestions, expressing concerns about the interpretability of AI outputs and workflow interruptions. Physician-AI collaboration scores showed moderate trust in the AI's suggestions (mean 3.7/5) but rated efficiency (2.0/5), ease of use (1.7/5), and workload reduction (1.7/5) as poor, indicating usability challenges and workflow interruptions.</p><p><strong>Discussion: </strong>In this single-center, randomized trial, AI-assisted EDX interpretation did not demonstrate a significant advantage over conventional physician-only interpretation. Nevertheless, the AI framework may help reduce workload and documentation burdens by handling simpler, routine EDX tests freeing physicians to focus on more complex cases that require greater expertise.</p><p><strong>Trial registration: </strong>ClinicalTrials.gov Identifier: NCT06902675.</p>","PeriodicalId":16558,"journal":{"name":"Journal of Neurology","volume":"272 9","pages":"586"},"PeriodicalIF":4.6000,"publicationDate":"2025-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12373542/pdf/","citationCount":"0","resultStr":"{\"title\":\"AI-Based EMG Reporting: A Randomized Controlled Trial.\",\"authors\":\"Alon Gorenshtein, Yana Weisblat, Mohamed Khateb, Gilad Kenan, Irina Tsirkin, Galina Fayn, Semion Geller, Shahar Shelly\",\"doi\":\"10.1007/s00415-025-13261-3\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Background and objectives: </strong>Accurate interpretation of electrodiagnostic (EDX) studies is essential for the diagnosis and management of neuromuscular disorders. Artificial intelligence (AI) based tools may improve consistency and quality of EDX reporting and reduce workload. The aim of this study is to evaluate the performance of an AI-assisted, multi-agent framework (INSPIRE) in comparison with standard physician interpretation in a randomized controlled trial (RCT).</p><p><strong>Methods: </strong>We prospectively enrolled 200 patients (out of 363 assessed for eligibility) referred for EDX. Patients were randomly assigned to either a control group (physician-only interpretation) or an intervention group (physician-AI interpretation). Three board-certified physicians, rotated across both arms. In the intervention group, an AI-generated preliminary report was combined with the physician's independent findings (human-AI integration). The primary outcome was EDX report quality using score we developed named AI-Generated EMG Report Score (AIGERS; score range, 0-1, with higher scores indicating more accurate or complete reports). Secondary outcomes included a physician-reported AI integration rating score (PAIR) and a compliance survey evaluating ease of AI adoption.</p><p><strong>Results: </strong>Of the 200 enrolled patients, 100 were allocated to AI-assisted interpretation and 100 to physician-only reporting. While AI-generated preliminary reports offered moderate consistency on the AIGERS metric, the integrated (physician-AI) approach did not significantly outperform physician-only methods. Despite some anecdotal advantages such as efficiency in suggesting standardized terminology quantitatively, the AIGERS scores for physician-AI integration was nearly the same as those in the physician-only arm and did not reach statistical significance (p > 0.05 for all comparisons). Physicians reported variable acceptance of AI suggestions, expressing concerns about the interpretability of AI outputs and workflow interruptions. Physician-AI collaboration scores showed moderate trust in the AI's suggestions (mean 3.7/5) but rated efficiency (2.0/5), ease of use (1.7/5), and workload reduction (1.7/5) as poor, indicating usability challenges and workflow interruptions.</p><p><strong>Discussion: </strong>In this single-center, randomized trial, AI-assisted EDX interpretation did not demonstrate a significant advantage over conventional physician-only interpretation. Nevertheless, the AI framework may help reduce workload and documentation burdens by handling simpler, routine EDX tests freeing physicians to focus on more complex cases that require greater expertise.</p><p><strong>Trial registration: </strong>ClinicalTrials.gov Identifier: NCT06902675.</p>\",\"PeriodicalId\":16558,\"journal\":{\"name\":\"Journal of Neurology\",\"volume\":\"272 9\",\"pages\":\"586\"},\"PeriodicalIF\":4.6000,\"publicationDate\":\"2025-08-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12373542/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Neurology\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1007/s00415-025-13261-3\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"CLINICAL NEUROLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Neurology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1007/s00415-025-13261-3","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"CLINICAL NEUROLOGY","Score":null,"Total":0}
AI-Based EMG Reporting: A Randomized Controlled Trial.
Background and objectives: Accurate interpretation of electrodiagnostic (EDX) studies is essential for the diagnosis and management of neuromuscular disorders. Artificial intelligence (AI) based tools may improve consistency and quality of EDX reporting and reduce workload. The aim of this study is to evaluate the performance of an AI-assisted, multi-agent framework (INSPIRE) in comparison with standard physician interpretation in a randomized controlled trial (RCT).
Methods: We prospectively enrolled 200 patients (out of 363 assessed for eligibility) referred for EDX. Patients were randomly assigned to either a control group (physician-only interpretation) or an intervention group (physician-AI interpretation). Three board-certified physicians, rotated across both arms. In the intervention group, an AI-generated preliminary report was combined with the physician's independent findings (human-AI integration). The primary outcome was EDX report quality using score we developed named AI-Generated EMG Report Score (AIGERS; score range, 0-1, with higher scores indicating more accurate or complete reports). Secondary outcomes included a physician-reported AI integration rating score (PAIR) and a compliance survey evaluating ease of AI adoption.
Results: Of the 200 enrolled patients, 100 were allocated to AI-assisted interpretation and 100 to physician-only reporting. While AI-generated preliminary reports offered moderate consistency on the AIGERS metric, the integrated (physician-AI) approach did not significantly outperform physician-only methods. Despite some anecdotal advantages such as efficiency in suggesting standardized terminology quantitatively, the AIGERS scores for physician-AI integration was nearly the same as those in the physician-only arm and did not reach statistical significance (p > 0.05 for all comparisons). Physicians reported variable acceptance of AI suggestions, expressing concerns about the interpretability of AI outputs and workflow interruptions. Physician-AI collaboration scores showed moderate trust in the AI's suggestions (mean 3.7/5) but rated efficiency (2.0/5), ease of use (1.7/5), and workload reduction (1.7/5) as poor, indicating usability challenges and workflow interruptions.
Discussion: In this single-center, randomized trial, AI-assisted EDX interpretation did not demonstrate a significant advantage over conventional physician-only interpretation. Nevertheless, the AI framework may help reduce workload and documentation burdens by handling simpler, routine EDX tests freeing physicians to focus on more complex cases that require greater expertise.
期刊介绍:
The Journal of Neurology is an international peer-reviewed journal which provides a source for publishing original communications and reviews on clinical neurology covering the whole field.
In addition, Letters to the Editors serve as a forum for clinical cases and the exchange of ideas which highlight important new findings. A section on Neurological progress serves to summarise the major findings in certain fields of neurology. Commentaries on new developments in clinical neuroscience, which may be commissioned or submitted, are published as editorials.
Every neurologist interested in the current diagnosis and treatment of neurological disorders needs access to the information contained in this valuable journal.