{"title":"DCEF-AVNet: multi-scale feature fusion and attention mechanism-guided brain tumor segmentation network.","authors":"Linlin Wang, Tong Zhang, Chuanyun Wang, Qian Gao, Zhongyi Li, Jing Shao","doi":"10.1117/1.JMI.12.2.024503","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>Accurate and efficient automatic segmentation of brain tumors is critical for diagnosis and treatment. However, the diversity in the appearance, location, and shape of brain tumors and their subregions, coupled with complex boundaries, presents significant challenges. We aim to improve segmentation accuracy by addressing limitations in V-Net, including insufficient utilization of multi-scale features and difficulties in managing complex spatial relationships and long-range dependencies.</p><p><strong>Approach: </strong>We propose an improved network structure, dynamic convolution enhanced fusion axial V-Net (DCEF-AVNet), which integrates an enhanced feature fusion module and axial attention mechanisms. The feature fusion module integrates dynamic convolution with a redesigned skip connection strategy to effectively combine multi-scale features, reducing feature inconsistencies and improving representation capability. Axial attention mechanisms are introduced at encoder-decoder connections to manage spatial relationships and alleviate long-range dependency issues. The network was evaluated using the BraTS2021 dataset, with performance measured in terms of Dice coefficients and Hausdorff distances.</p><p><strong>Results: </strong>DCEF-AVNet achieved Dice coefficients of 92.49%, 91.35%, and 91.96% for the whole tumor (WT), tumor core (TC), and enhancing tumor (ET) regions, respectively, significantly outperforming baseline methods. The model also demonstrated robust performance across multiple runs, with consistently low standard deviations in metrics.</p><p><strong>Conclusions: </strong>The integration of dynamic convolution, enhanced feature fusion, and axial attention mechanisms enables DCEF-AVNet to deliver superior segmentation accuracy and robustness. These results underscore its potential for advancing automated brain tumor segmentation and improving clinical decision-making.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 2","pages":"024503"},"PeriodicalIF":1.9000,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11925075/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Medical Imaging","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1117/1.JMI.12.2.024503","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/3/20 0:00:00","PubModel":"Epub","JCR":"Q3","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0
Abstract
Purpose: Accurate and efficient automatic segmentation of brain tumors is critical for diagnosis and treatment. However, the diversity in the appearance, location, and shape of brain tumors and their subregions, coupled with complex boundaries, presents significant challenges. We aim to improve segmentation accuracy by addressing limitations in V-Net, including insufficient utilization of multi-scale features and difficulties in managing complex spatial relationships and long-range dependencies.
Approach: We propose an improved network structure, dynamic convolution enhanced fusion axial V-Net (DCEF-AVNet), which integrates an enhanced feature fusion module and axial attention mechanisms. The feature fusion module integrates dynamic convolution with a redesigned skip connection strategy to effectively combine multi-scale features, reducing feature inconsistencies and improving representation capability. Axial attention mechanisms are introduced at encoder-decoder connections to manage spatial relationships and alleviate long-range dependency issues. The network was evaluated using the BraTS2021 dataset, with performance measured in terms of Dice coefficients and Hausdorff distances.
Results: DCEF-AVNet achieved Dice coefficients of 92.49%, 91.35%, and 91.96% for the whole tumor (WT), tumor core (TC), and enhancing tumor (ET) regions, respectively, significantly outperforming baseline methods. The model also demonstrated robust performance across multiple runs, with consistently low standard deviations in metrics.
Conclusions: The integration of dynamic convolution, enhanced feature fusion, and axial attention mechanisms enables DCEF-AVNet to deliver superior segmentation accuracy and robustness. These results underscore its potential for advancing automated brain tumor segmentation and improving clinical decision-making.
期刊介绍:
JMI covers fundamental and translational research, as well as applications, focused on medical imaging, which continue to yield physical and biomedical advancements in the early detection, diagnostics, and therapy of disease as well as in the understanding of normal. The scope of JMI includes: Imaging physics, Tomographic reconstruction algorithms (such as those in CT and MRI), Image processing and deep learning, Computer-aided diagnosis and quantitative image analysis, Visualization and modeling, Picture archiving and communications systems (PACS), Image perception and observer performance, Technology assessment, Ultrasonic imaging, Image-guided procedures, Digital pathology, Biomedical applications of biomedical imaging. JMI allows for the peer-reviewed communication and archiving of scientific developments, translational and clinical applications, reviews, and recommendations for the field.