Zihan Zhao , Nguyen Quoc Khanh Le , Matthew Chin Heng Chua
{"title":"AI-driven multi-modal framework for prognostic modeling in glioblastoma: Enhancing clinical decision support","authors":"Zihan Zhao , Nguyen Quoc Khanh Le , Matthew Chin Heng Chua","doi":"10.1016/j.compmedimag.2025.102628","DOIUrl":null,"url":null,"abstract":"<div><h3>Objective</h3><div>Glioblastoma (GBM) is the most aggressive malignant brain tumor, associated with poor prognosis and limited therapeutic options. Accurate prognostic modeling is essential for guiding personalized treatment strategies. However, existing models often rely on single-modality data, limiting their ability to capture the complex molecular and histopathological heterogeneity of GBM. This study proposes an AI-driven, multi-modal framework for clinical decision support, encompassing both early triage and prognostic evaluation stages. A Vision Transformer (ViT) is first employed to classify tumor grades (WHO grades 2–4) using radiological images. Subsequently, an attention-based deep learning model integrates histopathological and transcriptomic data to improve risk stratification and inform treatment planning.</div></div><div><h3>Methods</h3><div>The ViT model was trained on FLAIR MRI scans from the UCSF-PDGM dataset to perform tumor grading during the early triage phase. For prognostic modeling, whole-slide histopathological images and RNA sequencing profiles were integrated using an attention-based deep learning architecture. Model performance was evaluated using the area under the curve (AUC), concordance index (C-index), and Kaplan–Meier survival analysis across independent CPTAC-GBM and TCGA-GBM cohorts.</div></div><div><h3>Results</h3><div>The ViT model achieved F1-scores exceeding 0.89 across all WHO tumor grades. The multi-modal model significantly outperformed single-modality baselines, demonstrating higher C-index values and superior prognostic accuracy. Kaplan–Meier analysis revealed statistically significant differences (p < 0.0001) between high- and low-risk patient groups.</div></div><div><h3>Conclusion</h3><div>This AI-enabled, multi-modal framework improves clinical decision support in GBM by enabling accurate risk stratification and treatment planning. The integration of radiological imaging, histopathology, and transcriptomics offers a comprehensive and personalized approach to GBM prognosis.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"124 ","pages":"Article 102628"},"PeriodicalIF":4.9000,"publicationDate":"2025-08-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computerized Medical Imaging and Graphics","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0895611125001375","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Objective
Glioblastoma (GBM) is the most aggressive malignant brain tumor, associated with poor prognosis and limited therapeutic options. Accurate prognostic modeling is essential for guiding personalized treatment strategies. However, existing models often rely on single-modality data, limiting their ability to capture the complex molecular and histopathological heterogeneity of GBM. This study proposes an AI-driven, multi-modal framework for clinical decision support, encompassing both early triage and prognostic evaluation stages. A Vision Transformer (ViT) is first employed to classify tumor grades (WHO grades 2–4) using radiological images. Subsequently, an attention-based deep learning model integrates histopathological and transcriptomic data to improve risk stratification and inform treatment planning.
Methods
The ViT model was trained on FLAIR MRI scans from the UCSF-PDGM dataset to perform tumor grading during the early triage phase. For prognostic modeling, whole-slide histopathological images and RNA sequencing profiles were integrated using an attention-based deep learning architecture. Model performance was evaluated using the area under the curve (AUC), concordance index (C-index), and Kaplan–Meier survival analysis across independent CPTAC-GBM and TCGA-GBM cohorts.
Results
The ViT model achieved F1-scores exceeding 0.89 across all WHO tumor grades. The multi-modal model significantly outperformed single-modality baselines, demonstrating higher C-index values and superior prognostic accuracy. Kaplan–Meier analysis revealed statistically significant differences (p < 0.0001) between high- and low-risk patient groups.
Conclusion
This AI-enabled, multi-modal framework improves clinical decision support in GBM by enabling accurate risk stratification and treatment planning. The integration of radiological imaging, histopathology, and transcriptomics offers a comprehensive and personalized approach to GBM prognosis.
期刊介绍:
The purpose of the journal Computerized Medical Imaging and Graphics is to act as a source for the exchange of research results concerning algorithmic advances, development, and application of digital imaging in disease detection, diagnosis, intervention, prevention, precision medicine, and population health. Included in the journal will be articles on novel computerized imaging or visualization techniques, including artificial intelligence and machine learning, augmented reality for surgical planning and guidance, big biomedical data visualization, computer-aided diagnosis, computerized-robotic surgery, image-guided therapy, imaging scanning and reconstruction, mobile and tele-imaging, radiomics, and imaging integration and modeling with other information relevant to digital health. The types of biomedical imaging include: magnetic resonance, computed tomography, ultrasound, nuclear medicine, X-ray, microwave, optical and multi-photon microscopy, video and sensory imaging, and the convergence of biomedical images with other non-imaging datasets.