Bgman: Boundary-Prior-Guided Multi-scale Aggregation Network for skin lesion segmentation

IF 3.1 3区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Zhenyang Huang, Yixing Zhao, Jinjiang Li, Yepeng Liu
{"title":"Bgman: Boundary-Prior-Guided Multi-scale Aggregation Network for skin lesion segmentation","authors":"Zhenyang Huang, Yixing Zhao, Jinjiang Li, Yepeng Liu","doi":"10.1007/s13042-024-02284-3","DOIUrl":null,"url":null,"abstract":"<p>Skin lesion segmentation is a fundamental task in the field of medical image analysis. Deep learning approaches have become essential tools for segmenting medical images, as their accuracy in effectively analyzing abnormalities plays a critical role in determining the ultimate diagnostic results. Because of the inherent difficulties presented by medical images, including variations in shapes and sizes, along with the indistinct boundaries between lesions and the surrounding backgrounds, certain conventional algorithms face difficulties in fulfilling the growing requirements for elevated accuracy in processing medical images. To enhance the performance in capturing edge features and fine details of lesion processing, this paper presents the Boundary-Prior-Guided Multi-Scale Aggregation Network for skin lesion segmentation (BGMAN). The proposed BGMAN follows a basic Encoder–Decoder structure, wherein the encoder network employs prevalent CNN-based architectures to capture semantic information. We propose the Transformer Bridge Block (TBB) and employ it to enhance multi-scale features captured by the encoder. The TBB strengthens the intensity of weak feature information, establishing long-distance relationships between feature information. In order to augment BGMAN’s capability to identify boundaries, a boundary-guided decoder is designed, utilizing the Boundary Aware Block (BAB) and Cross Scale Fusion Block (CSFB) to guide the decoding learning process. BAB can acquire features embedded with explicit boundary information under the supervision of a boundary mask, while CSFB aggregates boundary features from different scales using learnable embeddings. The proposed method has been validated on the ISIC2016, ISIC2017, and <span>\\(PH^2\\)</span> datasets. It outperforms current mainstream networks with the following results: F1 92.99 and IoU 87.71 on ISIC2016, F1 86.42 and IoU 78.34 on ISIC2017, and F1 94.83 and IoU 90.26 on <span>\\(PH^2\\)</span>.</p>","PeriodicalId":51327,"journal":{"name":"International Journal of Machine Learning and Cybernetics","volume":"43 1","pages":""},"PeriodicalIF":3.1000,"publicationDate":"2024-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Machine Learning and Cybernetics","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s13042-024-02284-3","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Skin lesion segmentation is a fundamental task in the field of medical image analysis. Deep learning approaches have become essential tools for segmenting medical images, as their accuracy in effectively analyzing abnormalities plays a critical role in determining the ultimate diagnostic results. Because of the inherent difficulties presented by medical images, including variations in shapes and sizes, along with the indistinct boundaries between lesions and the surrounding backgrounds, certain conventional algorithms face difficulties in fulfilling the growing requirements for elevated accuracy in processing medical images. To enhance the performance in capturing edge features and fine details of lesion processing, this paper presents the Boundary-Prior-Guided Multi-Scale Aggregation Network for skin lesion segmentation (BGMAN). The proposed BGMAN follows a basic Encoder–Decoder structure, wherein the encoder network employs prevalent CNN-based architectures to capture semantic information. We propose the Transformer Bridge Block (TBB) and employ it to enhance multi-scale features captured by the encoder. The TBB strengthens the intensity of weak feature information, establishing long-distance relationships between feature information. In order to augment BGMAN’s capability to identify boundaries, a boundary-guided decoder is designed, utilizing the Boundary Aware Block (BAB) and Cross Scale Fusion Block (CSFB) to guide the decoding learning process. BAB can acquire features embedded with explicit boundary information under the supervision of a boundary mask, while CSFB aggregates boundary features from different scales using learnable embeddings. The proposed method has been validated on the ISIC2016, ISIC2017, and \(PH^2\) datasets. It outperforms current mainstream networks with the following results: F1 92.99 and IoU 87.71 on ISIC2016, F1 86.42 and IoU 78.34 on ISIC2017, and F1 94.83 and IoU 90.26 on \(PH^2\).

Abstract Image

Bgman:用于皮损分割的边界先导多尺度聚合网络
皮肤病变分割是医学图像分析领域的一项基本任务。深度学习方法已成为分割医学图像的重要工具,因为它们在有效分析异常情况方面的准确性对最终诊断结果起着至关重要的作用。由于医学图像本身存在的困难,包括形状和大小的变化,以及病变和周围背景之间界限不清,某些传统算法难以满足对提高医学图像处理准确性日益增长的要求。为了提高捕捉边缘特征和病变处理细节的性能,本文提出了用于皮肤病变分割的边界先导多尺度聚合网络(BGMAN)。所提出的 BGMAN 遵循基本的编码器-解码器结构,其中编码器网络采用流行的基于 CNN 的架构来捕捉语义信息。我们提出了变换器桥块(TBB),并利用它来增强编码器捕捉到的多尺度特征。TBB 可增强弱特征信息的强度,建立特征信息之间的远距离关系。为了增强 BGMAN 识别边界的能力,我们设计了一个边界引导解码器,利用边界感知块(BAB)和跨尺度融合块(CSFB)来引导解码学习过程。BAB 可以在边界掩码的监督下获取嵌入了明确边界信息的特征,而 CSFB 则利用可学习的嵌入来聚合来自不同尺度的边界特征。所提出的方法在 ISIC2016、ISIC2017 和 (PH^2\)数据集上得到了验证。其结果如下,优于当前的主流网络:在 ISIC2016 上,F1 为 92.99,IoU 为 87.71;在 ISIC2017 上,F1 为 86.42,IoU 为 78.34;在 \(PH^2\) 上,F1 为 94.83,IoU 为 90.26。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
International Journal of Machine Learning and Cybernetics
International Journal of Machine Learning and Cybernetics COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE-
CiteScore
7.90
自引率
10.70%
发文量
225
期刊介绍: Cybernetics is concerned with describing complex interactions and interrelationships between systems which are omnipresent in our daily life. Machine Learning discovers fundamental functional relationships between variables and ensembles of variables in systems. The merging of the disciplines of Machine Learning and Cybernetics is aimed at the discovery of various forms of interaction between systems through diverse mechanisms of learning from data. The International Journal of Machine Learning and Cybernetics (IJMLC) focuses on the key research problems emerging at the junction of machine learning and cybernetics and serves as a broad forum for rapid dissemination of the latest advancements in the area. The emphasis of IJMLC is on the hybrid development of machine learning and cybernetics schemes inspired by different contributing disciplines such as engineering, mathematics, cognitive sciences, and applications. New ideas, design alternatives, implementations and case studies pertaining to all the aspects of machine learning and cybernetics fall within the scope of the IJMLC. Key research areas to be covered by the journal include: Machine Learning for modeling interactions between systems Pattern Recognition technology to support discovery of system-environment interaction Control of system-environment interactions Biochemical interaction in biological and biologically-inspired systems Learning for improvement of communication schemes between systems
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信