{"title":"An attention based residual U-Net with swin transformer for brain MRI segmentation","authors":"Tazkia Mim Angona, M. Rubaiyat Hossain Mondal","doi":"10.1016/j.array.2025.100376","DOIUrl":null,"url":null,"abstract":"<div><div>Brain Tumors are a life-threatening cancer type. Due to the varied types and aggressive nature of these tumors, medical diagnostics faces significant challenges. Effective diagnosis and treatment planning depends on identifying the brain tumor areas from MRI images accurately. Traditional methods tend to use manual segmentation, which is costly, time consuming and prone to errors. Automated segmentation using deep learning approaches has shown potential in detecting tumor region. However, the complexity of the tumor areas which contain various shapes, sizes, fuzzy boundaries, makes this process difficult. Therefore, a robust automated segmentation method in brain tumor segmentation is required. In our paper, we present a hybrid model, 3-Dimension (3D) ResAttU-Net-Swin, which combines residual U-Net, attention mechanism and swin transformer. Residual blocks are introduced in the U-Net structure as encoder and decoder to avoid vanishing gradient problems and improve feature recovery. Attention-based skip connections are used to enhance the feature information transition between the encoder and decoder. The swin transformer obtains broad-scale features from the image data. The proposed hybrid model was evaluated on both the BraTS 2020 and BraTS 2019 datasets. It achieved an average Dice Similarity Coefficients (DSC) of 88.27 % and average Intersection over Union (IoU) of 79.93 % on BraTS 2020. On BraTS 2019, the model achieved an average DSC of 89.20 % and average IoU of 81.40 %. The model obtains higher DSC than the existing methods. The experiment result shows that the proposed methodology, 3D ResAttU-Net-Swin can be a potential for brain tumor segmentation in clinical settings.</div></div>","PeriodicalId":8417,"journal":{"name":"Array","volume":"25 ","pages":"Article 100376"},"PeriodicalIF":2.3000,"publicationDate":"2025-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Array","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2590005625000037","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0
Abstract
Brain Tumors are a life-threatening cancer type. Due to the varied types and aggressive nature of these tumors, medical diagnostics faces significant challenges. Effective diagnosis and treatment planning depends on identifying the brain tumor areas from MRI images accurately. Traditional methods tend to use manual segmentation, which is costly, time consuming and prone to errors. Automated segmentation using deep learning approaches has shown potential in detecting tumor region. However, the complexity of the tumor areas which contain various shapes, sizes, fuzzy boundaries, makes this process difficult. Therefore, a robust automated segmentation method in brain tumor segmentation is required. In our paper, we present a hybrid model, 3-Dimension (3D) ResAttU-Net-Swin, which combines residual U-Net, attention mechanism and swin transformer. Residual blocks are introduced in the U-Net structure as encoder and decoder to avoid vanishing gradient problems and improve feature recovery. Attention-based skip connections are used to enhance the feature information transition between the encoder and decoder. The swin transformer obtains broad-scale features from the image data. The proposed hybrid model was evaluated on both the BraTS 2020 and BraTS 2019 datasets. It achieved an average Dice Similarity Coefficients (DSC) of 88.27 % and average Intersection over Union (IoU) of 79.93 % on BraTS 2020. On BraTS 2019, the model achieved an average DSC of 89.20 % and average IoU of 81.40 %. The model obtains higher DSC than the existing methods. The experiment result shows that the proposed methodology, 3D ResAttU-Net-Swin can be a potential for brain tumor segmentation in clinical settings.