Gian Marco Conte, Mana Moassefi, Paul A Decker, Matthew L Kosel, Christina B McCarthy, Jessica A Sagen, Yalda Nikanpour, Mahboubeh Fereidan-Esfahani, Michael W Ruff, Fiorella S Guido, Heather K Pump, Terry C Burns, Robert B Jenkins, Bradley J Erickson, Daniel H Lachance, W Oliver Tobin, Jeanette E Eckel-Passow
{"title":"Deep learning MRI models for the differential diagnosis of tumefactive demyelination versus <i>IDH</i>-wildtype glioblastoma.","authors":"Gian Marco Conte, Mana Moassefi, Paul A Decker, Matthew L Kosel, Christina B McCarthy, Jessica A Sagen, Yalda Nikanpour, Mahboubeh Fereidan-Esfahani, Michael W Ruff, Fiorella S Guido, Heather K Pump, Terry C Burns, Robert B Jenkins, Bradley J Erickson, Daniel H Lachance, W Oliver Tobin, Jeanette E Eckel-Passow","doi":"10.3174/ajnr.A8645","DOIUrl":null,"url":null,"abstract":"<p><strong>Background and purpose: </strong>Diagnosis of tumefactive demyelination can be challenging. The diagnosis of indeterminate brain lesions on MRI often requires tissue confirmation via brain biopsy. Noninvasive methods for accurate diagnosis of tumor and non-tumor etiologies allows for tailored therapy, optimal tumor control, and a reduced risk of iatrogenic morbidity and mortality. Tumefactive demyelination has imaging features that mimic <i>isocitrate dehydrogenase</i>-wildtype glioblastoma (<i>IDH</i>wt GBM). We hypothesized that deep learning applied to postcontrast T1-weighted (T1C) and T2-weighted (T2) MRI images can discriminate tumefactive demyelination from <i>IDH</i>wt GBM.</p><p><strong>Materials and methods: </strong>Patients with tumefactive demyelination (n=144) and <i>IDH</i>wt GBM (n=455) were identified by clinical registries. A 3D DenseNet121 architecture was used to develop models to differentiate tumefactive demyelination and <i>IDH</i>wt GBM using both T1C and T2 MRI images, as well as only T1C and only T2 images. A three-stage design was used: (i) model development and internal validation via five-fold cross validation using a sex-, age-, and MRI technology-matched set of tumefactive demyelination and <i>IDH</i>wt GBM, (ii) validation of model specificity on independent <i>IDH</i>wt GBM, and (iii) prospective validation on tumefactive demyelination and <i>IDH</i>wt GBM. Stratified AUCs were used to evaluate model performance stratified by sex, age at diagnosis, MRI scanner strength, and MRI acquisition.</p><p><strong>Results: </strong>The deep learning model developed using both T1C and T2 images had a prospective validation area under the receiver operator characteristic curve (AUC) of 88% (95% CI: 0.82 - 0.95). In the prospective validation stage, a model score threshold of 0.28 resulted in 91% sensitivity of correctly classifying tumefactive demyelination and 80% specificity (correctly classifying <i>IDH</i>wt GBM). Stratified AUCs demonstrated that model performance may be improved if thresholds were chosen stratified by age and MRI acquisition.</p><p><strong>Conclusions: </strong>MRI images can provide the basis for applying deep learning models to aid in the differential diagnosis of brain lesions. Further validation is needed to evaluate how well the model generalizes across institutions, patient populations, and technology, and to evaluate optimal thresholds for classification. Next steps also should incorporate additional tumor etiologies such as CNS lymphoma and brain metastases.</p><p><strong>Abbreviations: </strong>AUC = area under the receiver operator characteristic curve; CNS = central nervous system; CNSIDD = central nervous system inflammatory demyelinating disease; FeTS = federated tumor segmentation; GBM = glioblastoma; <i>IDH</i>wt = <i>isocitrate dehydrogenase</i> wildtype; IHC = immunohistochemistry; MOGAD = myelin oligodendrocyte glycoprotein antibody associated disorder; MS = multiple sclerosis; NMOSD = neuromyelitis optica spectrum disorder; wt = wildtype.</p>","PeriodicalId":93863,"journal":{"name":"AJNR. American journal of neuroradiology","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2025-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"AJNR. American journal of neuroradiology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3174/ajnr.A8645","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Background and purpose: Diagnosis of tumefactive demyelination can be challenging. The diagnosis of indeterminate brain lesions on MRI often requires tissue confirmation via brain biopsy. Noninvasive methods for accurate diagnosis of tumor and non-tumor etiologies allows for tailored therapy, optimal tumor control, and a reduced risk of iatrogenic morbidity and mortality. Tumefactive demyelination has imaging features that mimic isocitrate dehydrogenase-wildtype glioblastoma (IDHwt GBM). We hypothesized that deep learning applied to postcontrast T1-weighted (T1C) and T2-weighted (T2) MRI images can discriminate tumefactive demyelination from IDHwt GBM.
Materials and methods: Patients with tumefactive demyelination (n=144) and IDHwt GBM (n=455) were identified by clinical registries. A 3D DenseNet121 architecture was used to develop models to differentiate tumefactive demyelination and IDHwt GBM using both T1C and T2 MRI images, as well as only T1C and only T2 images. A three-stage design was used: (i) model development and internal validation via five-fold cross validation using a sex-, age-, and MRI technology-matched set of tumefactive demyelination and IDHwt GBM, (ii) validation of model specificity on independent IDHwt GBM, and (iii) prospective validation on tumefactive demyelination and IDHwt GBM. Stratified AUCs were used to evaluate model performance stratified by sex, age at diagnosis, MRI scanner strength, and MRI acquisition.
Results: The deep learning model developed using both T1C and T2 images had a prospective validation area under the receiver operator characteristic curve (AUC) of 88% (95% CI: 0.82 - 0.95). In the prospective validation stage, a model score threshold of 0.28 resulted in 91% sensitivity of correctly classifying tumefactive demyelination and 80% specificity (correctly classifying IDHwt GBM). Stratified AUCs demonstrated that model performance may be improved if thresholds were chosen stratified by age and MRI acquisition.
Conclusions: MRI images can provide the basis for applying deep learning models to aid in the differential diagnosis of brain lesions. Further validation is needed to evaluate how well the model generalizes across institutions, patient populations, and technology, and to evaluate optimal thresholds for classification. Next steps also should incorporate additional tumor etiologies such as CNS lymphoma and brain metastases.
Abbreviations: AUC = area under the receiver operator characteristic curve; CNS = central nervous system; CNSIDD = central nervous system inflammatory demyelinating disease; FeTS = federated tumor segmentation; GBM = glioblastoma; IDHwt = isocitrate dehydrogenase wildtype; IHC = immunohistochemistry; MOGAD = myelin oligodendrocyte glycoprotein antibody associated disorder; MS = multiple sclerosis; NMOSD = neuromyelitis optica spectrum disorder; wt = wildtype.