Decoupled pixel-wise correction for abdominal multi-organ segmentation

IF 5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Xiangchun Yu, Longjun Ding, Dingwen Zhang, Jianqing Wu, Miaomiao Liang, Jian Zheng, Wei Pang
{"title":"Decoupled pixel-wise correction for abdominal multi-organ segmentation","authors":"Xiangchun Yu, Longjun Ding, Dingwen Zhang, Jianqing Wu, Miaomiao Liang, Jian Zheng, Wei Pang","doi":"10.1007/s40747-025-01796-x","DOIUrl":null,"url":null,"abstract":"<p>The attention mechanism has emerged as a crucial component in medical image segmentation. Attention-based deep neural networks (ADNNs) fundamentally engage in the iterative computation of gradients for both input layers and weight parameters. Our research reveals a remarkable similarity between the optimization trajectory of ADNN and non-negative matrix factorization (NMF), where the latter involves the alternate adjustment of the base and coefficient matrices. This similarity implies that the alternating optimization strategy—characterized by the adjustment of input features by the attention mechanism and the adjustment of network weights—is central to the efficacy of attention mechanisms in ADNNs. Drawing an analogy to the NMF approach, we advocate for a pixel-wise adjustment of the input layer within ADNNs. Furthermore, to reduce the computational burden, we have developed a decoupled pixel-wise attention module (DPAM) and a self-attention module (DPSM). These modules are designed to counteract the challenges posed by the high inter-class similarity among different organs when performing multi-organ segmentation. The integration of our DPAM and DPSM into traditional network architectures facilitates the creation of an NMF-inspired ADNN framework, known as the DPC-Net, which comes in two variants: DPCA-Net for attention and DPCS-Net for self-attention. Our extensive experiments on the Synapse and FLARE22 datasets demonstrate that the DPC-Net achieves satisfactory performance and visualization results with lower computational cost. Specifically, the DPC-Net achieved a Dice score of 77.98% on the Synapse dataset and 87.04% on the FLARE22 dataset, while possessing merely 14.991 million parameters. Notably, our findings indicate that DPC-Net, when equipped with convolutional attention, surpasses those networks utilizing Transformer attention mechanisms on multi-organ segmentation tasks. Our code is available at https://github.com/605671435/DPC-Net.</p>","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"37 1","pages":""},"PeriodicalIF":5.0000,"publicationDate":"2025-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Complex & Intelligent Systems","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s40747-025-01796-x","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

The attention mechanism has emerged as a crucial component in medical image segmentation. Attention-based deep neural networks (ADNNs) fundamentally engage in the iterative computation of gradients for both input layers and weight parameters. Our research reveals a remarkable similarity between the optimization trajectory of ADNN and non-negative matrix factorization (NMF), where the latter involves the alternate adjustment of the base and coefficient matrices. This similarity implies that the alternating optimization strategy—characterized by the adjustment of input features by the attention mechanism and the adjustment of network weights—is central to the efficacy of attention mechanisms in ADNNs. Drawing an analogy to the NMF approach, we advocate for a pixel-wise adjustment of the input layer within ADNNs. Furthermore, to reduce the computational burden, we have developed a decoupled pixel-wise attention module (DPAM) and a self-attention module (DPSM). These modules are designed to counteract the challenges posed by the high inter-class similarity among different organs when performing multi-organ segmentation. The integration of our DPAM and DPSM into traditional network architectures facilitates the creation of an NMF-inspired ADNN framework, known as the DPC-Net, which comes in two variants: DPCA-Net for attention and DPCS-Net for self-attention. Our extensive experiments on the Synapse and FLARE22 datasets demonstrate that the DPC-Net achieves satisfactory performance and visualization results with lower computational cost. Specifically, the DPC-Net achieved a Dice score of 77.98% on the Synapse dataset and 87.04% on the FLARE22 dataset, while possessing merely 14.991 million parameters. Notably, our findings indicate that DPC-Net, when equipped with convolutional attention, surpasses those networks utilizing Transformer attention mechanisms on multi-organ segmentation tasks. Our code is available at https://github.com/605671435/DPC-Net.

求助全文
约1分钟内获得全文 求助全文
来源期刊
Complex & Intelligent Systems
Complex & Intelligent Systems COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE-
CiteScore
9.60
自引率
10.30%
发文量
297
期刊介绍: Complex & Intelligent Systems aims to provide a forum for presenting and discussing novel approaches, tools and techniques meant for attaining a cross-fertilization between the broad fields of complex systems, computational simulation, and intelligent analytics and visualization. The transdisciplinary research that the journal focuses on will expand the boundaries of our understanding by investigating the principles and processes that underlie many of the most profound problems facing society today.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信