Xue Linyan, Cao Jie, Zhou Kexuan, Chen Houquan, Qi Chaoyi, Yin Xiaosong, Wang Jianing, Yang Kun
{"title":"基于变分自编码器的多模态融合网络用于区分SCLC脑转移与NSCLC脑转移。","authors":"Xue Linyan, Cao Jie, Zhou Kexuan, Chen Houquan, Qi Chaoyi, Yin Xiaosong, Wang Jianing, Yang Kun","doi":"10.1002/mp.17816","DOIUrl":null,"url":null,"abstract":"<div>\n \n \n <section>\n \n <h3> Background</h3>\n \n <p>Distinguishing small cell lung cancer brain metastases from non-small cell lung cancer brain metastases in MRI sequence images is crucial for the accurate diagnosis and treatment of lung cancer brain metastases. Multi-MRI modalities provide complementary and comprehensive information, but efficiently merging these sequences to achieve modality complementarity is challenging due to redundant information within radiomic features and heterogeneity across different modalities.</p>\n </section>\n \n <section>\n \n <h3> Purpose</h3>\n \n <p>To address these challenges, we propose a novel multimodal fusion network, termed MFN-VAE, which utilizes a variational auto-encoder (VAE) to compress and aggregate radiomic features derived from MRI images.</p>\n </section>\n \n <section>\n \n <h3> Methods</h3>\n \n <p>Initially, we extract radiomic features from areas of interest in MRI images across T1WI, FLAIR, and DWI modalities. A VAE encoder is then constructed to project these multimodal features into a latent space, where they are decoded into reconstruction features using a decoder. The encoder-decoder network is trained to extract the underlying feature representation of each modality, capturing both the consistency and specificity of each domain.</p>\n </section>\n \n <section>\n \n <h3> Results</h3>\n \n <p>Experimental results on our collected dataset of lung cancer brain metastases demonstrate the encouraging performance of our proposed MFN-VAE. The method achieved a 0.888 accuracy and a 0.920 AUC (area under the curve), outperforming state-of-the-art methods across different modal combinations.</p>\n </section>\n \n <section>\n \n <h3> Conclusions</h3>\n \n <p>In this study, we introduce the MFN-VAE, a new multimodal fusion network for differentiating small cell from non-small cell lung cancer brain metastases. Tested on a private dataset, MFN-VAE demonstrated high accuracy (ACC: 0.888; AUC: 0.920), effectively distinguishing between small cell lung cancer brain metastases (SCLC) and non-small cell lung cancer (NSCLC). The SHapley Additive explanation (SHAP) method was used to enhance model interpretability, providing clinicians with a reliable diagnostic tool. Overall, MFN-VAE shows great potential in improving the diagnosis and treatment of lung cancer brain metastases.</p>\n </section>\n </div>","PeriodicalId":18384,"journal":{"name":"Medical physics","volume":"52 7","pages":""},"PeriodicalIF":3.2000,"publicationDate":"2025-05-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A multimodal fusion network based on variational autoencoder for distinguishing SCLC brain metastases from NSCLC brain metastases\",\"authors\":\"Xue Linyan, Cao Jie, Zhou Kexuan, Chen Houquan, Qi Chaoyi, Yin Xiaosong, Wang Jianing, Yang Kun\",\"doi\":\"10.1002/mp.17816\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n \\n \\n <section>\\n \\n <h3> Background</h3>\\n \\n <p>Distinguishing small cell lung cancer brain metastases from non-small cell lung cancer brain metastases in MRI sequence images is crucial for the accurate diagnosis and treatment of lung cancer brain metastases. Multi-MRI modalities provide complementary and comprehensive information, but efficiently merging these sequences to achieve modality complementarity is challenging due to redundant information within radiomic features and heterogeneity across different modalities.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Purpose</h3>\\n \\n <p>To address these challenges, we propose a novel multimodal fusion network, termed MFN-VAE, which utilizes a variational auto-encoder (VAE) to compress and aggregate radiomic features derived from MRI images.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Methods</h3>\\n \\n <p>Initially, we extract radiomic features from areas of interest in MRI images across T1WI, FLAIR, and DWI modalities. A VAE encoder is then constructed to project these multimodal features into a latent space, where they are decoded into reconstruction features using a decoder. The encoder-decoder network is trained to extract the underlying feature representation of each modality, capturing both the consistency and specificity of each domain.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Results</h3>\\n \\n <p>Experimental results on our collected dataset of lung cancer brain metastases demonstrate the encouraging performance of our proposed MFN-VAE. The method achieved a 0.888 accuracy and a 0.920 AUC (area under the curve), outperforming state-of-the-art methods across different modal combinations.</p>\\n </section>\\n \\n <section>\\n \\n <h3> Conclusions</h3>\\n \\n <p>In this study, we introduce the MFN-VAE, a new multimodal fusion network for differentiating small cell from non-small cell lung cancer brain metastases. Tested on a private dataset, MFN-VAE demonstrated high accuracy (ACC: 0.888; AUC: 0.920), effectively distinguishing between small cell lung cancer brain metastases (SCLC) and non-small cell lung cancer (NSCLC). The SHapley Additive explanation (SHAP) method was used to enhance model interpretability, providing clinicians with a reliable diagnostic tool. Overall, MFN-VAE shows great potential in improving the diagnosis and treatment of lung cancer brain metastases.</p>\\n </section>\\n </div>\",\"PeriodicalId\":18384,\"journal\":{\"name\":\"Medical physics\",\"volume\":\"52 7\",\"pages\":\"\"},\"PeriodicalIF\":3.2000,\"publicationDate\":\"2025-05-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Medical physics\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/mp.17816\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical physics","FirstCategoryId":"3","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/mp.17816","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
A multimodal fusion network based on variational autoencoder for distinguishing SCLC brain metastases from NSCLC brain metastases
Background
Distinguishing small cell lung cancer brain metastases from non-small cell lung cancer brain metastases in MRI sequence images is crucial for the accurate diagnosis and treatment of lung cancer brain metastases. Multi-MRI modalities provide complementary and comprehensive information, but efficiently merging these sequences to achieve modality complementarity is challenging due to redundant information within radiomic features and heterogeneity across different modalities.
Purpose
To address these challenges, we propose a novel multimodal fusion network, termed MFN-VAE, which utilizes a variational auto-encoder (VAE) to compress and aggregate radiomic features derived from MRI images.
Methods
Initially, we extract radiomic features from areas of interest in MRI images across T1WI, FLAIR, and DWI modalities. A VAE encoder is then constructed to project these multimodal features into a latent space, where they are decoded into reconstruction features using a decoder. The encoder-decoder network is trained to extract the underlying feature representation of each modality, capturing both the consistency and specificity of each domain.
Results
Experimental results on our collected dataset of lung cancer brain metastases demonstrate the encouraging performance of our proposed MFN-VAE. The method achieved a 0.888 accuracy and a 0.920 AUC (area under the curve), outperforming state-of-the-art methods across different modal combinations.
Conclusions
In this study, we introduce the MFN-VAE, a new multimodal fusion network for differentiating small cell from non-small cell lung cancer brain metastases. Tested on a private dataset, MFN-VAE demonstrated high accuracy (ACC: 0.888; AUC: 0.920), effectively distinguishing between small cell lung cancer brain metastases (SCLC) and non-small cell lung cancer (NSCLC). The SHapley Additive explanation (SHAP) method was used to enhance model interpretability, providing clinicians with a reliable diagnostic tool. Overall, MFN-VAE shows great potential in improving the diagnosis and treatment of lung cancer brain metastases.
期刊介绍:
Medical Physics publishes original, high impact physics, imaging science, and engineering research that advances patient diagnosis and therapy through contributions in 1) Basic science developments with high potential for clinical translation 2) Clinical applications of cutting edge engineering and physics innovations 3) Broadly applicable and innovative clinical physics developments
Medical Physics is a journal of global scope and reach. By publishing in Medical Physics your research will reach an international, multidisciplinary audience including practicing medical physicists as well as physics- and engineering based translational scientists. We work closely with authors of promising articles to improve their quality.