Yuli Chen , Jiayang Bai , Jinjie Wang , Guoping Chen , Xinxin Zhang , Duan-Bo Shi , Xiujuan Lei , Peng Gao , Cheng Lu
{"title":"MSFusion: A multi-source hybrid feature fusion network for accurate grading of invasive breast cancer using H&E-stained histopathological images","authors":"Yuli Chen , Jiayang Bai , Jinjie Wang , Guoping Chen , Xinxin Zhang , Duan-Bo Shi , Xiujuan Lei , Peng Gao , Cheng Lu","doi":"10.1016/j.media.2025.103633","DOIUrl":null,"url":null,"abstract":"<div><div>Invasive breast cancer (IBC) is a prevalent malignant tumor in women, and precise grading plays a pivotal role in ensuring effective treatment and enhancing survival rates. However, accurately grading IBC presents a significant challenge due to its heterogeneous nature and the need to harness the complementary information from multiple nuclei sources in histopathology images. To tackle this critical problem, we introduce a novel multi-source hybrid feature fusion network named MSFusion. This network incorporates two types of hybrid features: deep learning features extracted by a novel Swin Transformer-based multi-branch network called MSwinT, and traditional handcrafted features that capture the morphological characteristics of multi-source nuclei. The primary branch of MSwinT captures the overall characteristics of the original images, while multiple auxiliary branches focus on identifying morphological features from diverse sources of nuclei, including tumor, mitotic, tubular, and epithelial nuclei. At each of the four stages for the branches in MSwinT, a functional KDC (key diagnostic components) fusion block with channel and spatial attentions is proposed to integrate the features extracted by all the branches. Ultimately, we synthesize the multi-source hybrid deep learning features and handcrafted features to improve the accuracy of IBC diagnosis and grading. Our multi-branch MSFusion network is rigorously evaluated on three distinct datasets, including two private clinical datasets (Qilu dataset and QDUH&SHSU dataset) as well as a publicly available Databiox dataset. The experimental results consistently demonstrate that our proposed MSFusion model outperforms the state-of-the-art methods. Specifically, the AUC for the Qilu dataset and QDUH&SHSU dataset are 81.3% and 90.2%, respectively, while the public Databiox dataset yields an AUC of 82.1%.</div></div>","PeriodicalId":18328,"journal":{"name":"Medical image analysis","volume":"104 ","pages":"Article 103633"},"PeriodicalIF":10.7000,"publicationDate":"2025-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Medical image analysis","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S136184152500180X","RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Invasive breast cancer (IBC) is a prevalent malignant tumor in women, and precise grading plays a pivotal role in ensuring effective treatment and enhancing survival rates. However, accurately grading IBC presents a significant challenge due to its heterogeneous nature and the need to harness the complementary information from multiple nuclei sources in histopathology images. To tackle this critical problem, we introduce a novel multi-source hybrid feature fusion network named MSFusion. This network incorporates two types of hybrid features: deep learning features extracted by a novel Swin Transformer-based multi-branch network called MSwinT, and traditional handcrafted features that capture the morphological characteristics of multi-source nuclei. The primary branch of MSwinT captures the overall characteristics of the original images, while multiple auxiliary branches focus on identifying morphological features from diverse sources of nuclei, including tumor, mitotic, tubular, and epithelial nuclei. At each of the four stages for the branches in MSwinT, a functional KDC (key diagnostic components) fusion block with channel and spatial attentions is proposed to integrate the features extracted by all the branches. Ultimately, we synthesize the multi-source hybrid deep learning features and handcrafted features to improve the accuracy of IBC diagnosis and grading. Our multi-branch MSFusion network is rigorously evaluated on three distinct datasets, including two private clinical datasets (Qilu dataset and QDUH&SHSU dataset) as well as a publicly available Databiox dataset. The experimental results consistently demonstrate that our proposed MSFusion model outperforms the state-of-the-art methods. Specifically, the AUC for the Qilu dataset and QDUH&SHSU dataset are 81.3% and 90.2%, respectively, while the public Databiox dataset yields an AUC of 82.1%.
浸润性乳腺癌(Invasive breast cancer, IBC)是女性常见的恶性肿瘤,准确的分级对保证有效治疗和提高生存率起着至关重要的作用。然而,由于IBC的异质性和需要利用组织病理学图像中多个核源的互补信息,准确分级IBC提出了一个重大挑战。为了解决这一关键问题,我们引入了一种新的多源混合特征融合网络MSFusion。该网络结合了两种类型的混合特征:一种是由基于Swin变压器的新型多分支网络MSwinT提取的深度学习特征,另一种是捕获多源核形态特征的传统手工特征。MSwinT的初级分支捕捉原始图像的总体特征,而多个辅助分支专注于识别来自不同核源的形态学特征,包括肿瘤核、有丝分裂核、小管核和上皮核。在MSwinT分支的每个阶段,提出了一个具有通道和空间关注的功能性KDC(关键诊断组件)融合块,以整合所有分支提取的特征。最后,我们综合了多源混合深度学习特征和手工特征,以提高IBC诊断和分级的准确性。我们的多分支MSFusion网络在三个不同的数据集上进行了严格的评估,包括两个私人临床数据集(Qilu数据集和qduh&shsu数据集)以及一个公开可用的Databiox数据集。实验结果一致表明,我们提出的MSFusion模型优于最先进的方法。其中,齐鲁数据集的AUC为81.3%,QDUH&;SHSU数据集的AUC为90.2%,而公共Databiox数据集的AUC为82.1%。
期刊介绍:
Medical Image Analysis serves as a platform for sharing new research findings in the realm of medical and biological image analysis, with a focus on applications of computer vision, virtual reality, and robotics to biomedical imaging challenges. The journal prioritizes the publication of high-quality, original papers contributing to the fundamental science of processing, analyzing, and utilizing medical and biological images. It welcomes approaches utilizing biomedical image datasets across all spatial scales, from molecular/cellular imaging to tissue/organ imaging.