GAIR-U-Net:利用多模态 MRI 图像进行脑肿瘤分割的三维引导注意残差 U 网

IF 5.2 2区 计算机科学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS
Evans Kipkoech Rutoh , Qin Zhi Guang , Noor Bahadar , Rehan Raza , Muhammad Shehzad Hanif
{"title":"GAIR-U-Net:利用多模态 MRI 图像进行脑肿瘤分割的三维引导注意残差 U 网","authors":"Evans Kipkoech Rutoh ,&nbsp;Qin Zhi Guang ,&nbsp;Noor Bahadar ,&nbsp;Rehan Raza ,&nbsp;Muhammad Shehzad Hanif","doi":"10.1016/j.jksuci.2024.102086","DOIUrl":null,"url":null,"abstract":"<div><p>Deep learning technologies have led to substantial breakthroughs in the field of biomedical image analysis. Accurate brain tumor segmentation is an essential aspect of treatment planning. Radiologists agree that manual segmentation is a difficult and time-consuming task that frequently causes delays in the diagnosing process. While U-Net-based methods have been widely used for brain tumor segmentation, many challenges persist, particularly when dealing with tumors of varying sizes, locations, and shapes. Additionally, segmenting tumor regions with structures requires a comprehensive model, which can increase computational complexity and potentially cause gradient vanishing issues. This study presents a novel method called 3D Guided Attention-based deep Inception Residual U-Net (GAIR-U-Net) to address these challenges. This model combines attention mechanisms, an inception module, and residual blocks with dilated convolution to enhance feature representation and spatial context understanding. The backbone of the model is the U-Net model, which leverages the power of inception and residual connections to capture intricate patterns and hierarchical features while expanding the model’s width in three-dimensional space without significantly increasing computational complexity. The attention mechanisms play a role in focusing on important regions and areas while downgrading irrelevant details. The dilated convolutions in the network help in learning both local and global information, improving accuracy and adaptability in segmenting tumors. All the experiments in this study were carried out on multimodal MRI scans that include (T1-weighted, T1-ce, T2-weighted, and FLAIR sequences) from the BraTS 2020 dataset. The presented model is trained and tested on the same dataset, which exhibited promising performance compared to previous methods. On the BraTS 2020 validation dataset, the proposed model obtained a dice score of 0.8796, 0.8634, and 0.8441 for whole tumor (WT), tumor core (TC), and enhancing tumor (ET), respectively. These results demonstrate the model’s efficacy in precisely segmenting brain tumors across various modalities. Comparative analyses underscore the model’s versatility in handling tumor shape variations, size, and location, making it a promising solution for clinical applications.</p></div>","PeriodicalId":48547,"journal":{"name":"Journal of King Saud University-Computer and Information Sciences","volume":null,"pages":null},"PeriodicalIF":5.2000,"publicationDate":"2024-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1319157824001757/pdfft?md5=8fbc647524401707d35ef8806d9bc8a7&pid=1-s2.0-S1319157824001757-main.pdf","citationCount":"0","resultStr":"{\"title\":\"GAIR-U-Net: 3D guided attention inception residual u-net for brain tumor segmentation using multimodal MRI images\",\"authors\":\"Evans Kipkoech Rutoh ,&nbsp;Qin Zhi Guang ,&nbsp;Noor Bahadar ,&nbsp;Rehan Raza ,&nbsp;Muhammad Shehzad Hanif\",\"doi\":\"10.1016/j.jksuci.2024.102086\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Deep learning technologies have led to substantial breakthroughs in the field of biomedical image analysis. Accurate brain tumor segmentation is an essential aspect of treatment planning. Radiologists agree that manual segmentation is a difficult and time-consuming task that frequently causes delays in the diagnosing process. While U-Net-based methods have been widely used for brain tumor segmentation, many challenges persist, particularly when dealing with tumors of varying sizes, locations, and shapes. Additionally, segmenting tumor regions with structures requires a comprehensive model, which can increase computational complexity and potentially cause gradient vanishing issues. This study presents a novel method called 3D Guided Attention-based deep Inception Residual U-Net (GAIR-U-Net) to address these challenges. This model combines attention mechanisms, an inception module, and residual blocks with dilated convolution to enhance feature representation and spatial context understanding. The backbone of the model is the U-Net model, which leverages the power of inception and residual connections to capture intricate patterns and hierarchical features while expanding the model’s width in three-dimensional space without significantly increasing computational complexity. The attention mechanisms play a role in focusing on important regions and areas while downgrading irrelevant details. The dilated convolutions in the network help in learning both local and global information, improving accuracy and adaptability in segmenting tumors. All the experiments in this study were carried out on multimodal MRI scans that include (T1-weighted, T1-ce, T2-weighted, and FLAIR sequences) from the BraTS 2020 dataset. The presented model is trained and tested on the same dataset, which exhibited promising performance compared to previous methods. On the BraTS 2020 validation dataset, the proposed model obtained a dice score of 0.8796, 0.8634, and 0.8441 for whole tumor (WT), tumor core (TC), and enhancing tumor (ET), respectively. These results demonstrate the model’s efficacy in precisely segmenting brain tumors across various modalities. Comparative analyses underscore the model’s versatility in handling tumor shape variations, size, and location, making it a promising solution for clinical applications.</p></div>\",\"PeriodicalId\":48547,\"journal\":{\"name\":\"Journal of King Saud University-Computer and Information Sciences\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":5.2000,\"publicationDate\":\"2024-06-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.sciencedirect.com/science/article/pii/S1319157824001757/pdfft?md5=8fbc647524401707d35ef8806d9bc8a7&pid=1-s2.0-S1319157824001757-main.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of King Saud University-Computer and Information Sciences\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1319157824001757\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of King Saud University-Computer and Information Sciences","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1319157824001757","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0

摘要

深度学习技术在生物医学图像分析领域取得了重大突破。准确的脑肿瘤分割是治疗计划的一个重要方面。放射科医生一致认为,人工分割是一项艰巨而耗时的任务,经常会延误诊断过程。虽然基于 U-Net 的方法已被广泛用于脑肿瘤分割,但仍存在许多挑战,尤其是在处理不同大小、位置和形状的肿瘤时。此外,分割具有结构的肿瘤区域需要一个全面的模型,这会增加计算复杂度,并可能导致梯度消失问题。本研究提出了一种名为 "基于 3D 引导注意力的深度初始残差 U-Net (GAIR-U-Net) "的新方法来应对这些挑战。该模型将注意力机制、初始模块和残差块与扩张卷积相结合,以增强特征表示和空间上下文理解。该模型的主干是 U-Net 模型,它利用起始和残差连接的力量来捕捉错综复杂的模式和层次特征,同时在不显著增加计算复杂度的情况下扩大模型在三维空间中的宽度。注意力机制的作用是聚焦重要区域和领域,同时降低无关细节的等级。网络中的扩张卷积有助于学习局部和全局信息,从而提高分割肿瘤的准确性和适应性。本研究的所有实验都是在 BraTS 2020 数据集中的多模态磁共振成像扫描(包括 T1 加权、T1-ce、T2 加权和 FLAIR 序列)上进行的。该模型在同一数据集上进行了训练和测试,与之前的方法相比,表现出了良好的性能。在 BraTS 2020 验证数据集上,所提出的模型在全肿瘤(WT)、肿瘤核心(TC)和增强肿瘤(ET)方面的骰子得分分别为 0.8796、0.8634 和 0.8441。这些结果证明了该模型在各种模式下精确分割脑肿瘤的功效。对比分析凸显了该模型在处理肿瘤形状变化、大小和位置方面的多功能性,使其成为临床应用的一个有前途的解决方案。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
GAIR-U-Net: 3D guided attention inception residual u-net for brain tumor segmentation using multimodal MRI images

Deep learning technologies have led to substantial breakthroughs in the field of biomedical image analysis. Accurate brain tumor segmentation is an essential aspect of treatment planning. Radiologists agree that manual segmentation is a difficult and time-consuming task that frequently causes delays in the diagnosing process. While U-Net-based methods have been widely used for brain tumor segmentation, many challenges persist, particularly when dealing with tumors of varying sizes, locations, and shapes. Additionally, segmenting tumor regions with structures requires a comprehensive model, which can increase computational complexity and potentially cause gradient vanishing issues. This study presents a novel method called 3D Guided Attention-based deep Inception Residual U-Net (GAIR-U-Net) to address these challenges. This model combines attention mechanisms, an inception module, and residual blocks with dilated convolution to enhance feature representation and spatial context understanding. The backbone of the model is the U-Net model, which leverages the power of inception and residual connections to capture intricate patterns and hierarchical features while expanding the model’s width in three-dimensional space without significantly increasing computational complexity. The attention mechanisms play a role in focusing on important regions and areas while downgrading irrelevant details. The dilated convolutions in the network help in learning both local and global information, improving accuracy and adaptability in segmenting tumors. All the experiments in this study were carried out on multimodal MRI scans that include (T1-weighted, T1-ce, T2-weighted, and FLAIR sequences) from the BraTS 2020 dataset. The presented model is trained and tested on the same dataset, which exhibited promising performance compared to previous methods. On the BraTS 2020 validation dataset, the proposed model obtained a dice score of 0.8796, 0.8634, and 0.8441 for whole tumor (WT), tumor core (TC), and enhancing tumor (ET), respectively. These results demonstrate the model’s efficacy in precisely segmenting brain tumors across various modalities. Comparative analyses underscore the model’s versatility in handling tumor shape variations, size, and location, making it a promising solution for clinical applications.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
10.50
自引率
8.70%
发文量
656
审稿时长
29 days
期刊介绍: In 2022 the Journal of King Saud University - Computer and Information Sciences will become an author paid open access journal. Authors who submit their manuscript after October 31st 2021 will be asked to pay an Article Processing Charge (APC) after acceptance of their paper to make their work immediately, permanently, and freely accessible to all. The Journal of King Saud University Computer and Information Sciences is a refereed, international journal that covers all aspects of both foundations of computer and its practical applications.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信