Ao Su , Xiaolin Wang , Hao Xu , Jianfeng Zhang , Kang Chen , Dexing Kong , Guangfei Li , Xiaojun Chen , Jianming Wen , Zhong Lv
{"title":"基于视觉感知的多任务学习多尺度乳腺癌超声图像分割与分类","authors":"Ao Su , Xiaolin Wang , Hao Xu , Jianfeng Zhang , Kang Chen , Dexing Kong , Guangfei Li , Xiaojun Chen , Jianming Wen , Zhong Lv","doi":"10.1016/j.bspc.2025.108212","DOIUrl":null,"url":null,"abstract":"<div><div>Breast tumor segmentation and classification are essential components of breast ultrasound (BUS) computer-aided diagnosis (CAD) systems, which help improve the accuracy of breast cancer diagnoses. However, challenges arise due to the complexity of tumor features, the intensity similarities between lesions and surrounding tissues, as well as variations in tumor shape and location. While deep learning has been widely applied in CAD systems, many existing approaches overlook the relationship between segmentation and classification tasks, limiting their effectiveness. We propose PMSMT-Net, a Perception-based Multi-scale Ultrasound Image Segmentation and Classification Multi-task Learning Network, which enhances both segmentation and classification performance in BUS images. The segmentation network integrates a Visual Perception Module (VPM) to simulate human-like focus on regions of interest and, combined with Multi-scale Dilated Convolution (MSDC), accurately captures morphological, locational, and edge features. To further improve segmentation accuracy, we introduce the Variable Residual Convolutional Block Attention Module (VR-CBAM) and the Receptive Field Block-based Perceptually Separable Convolution Module (RFB-PSC), which enhance context feature fusion and reduce spatial information loss. The VPM output, along with the segmentation results, is then used as input to the classification network, where a transfer learning and ensemble learning approach classifies breast tumors. Compared to state-of-the-art methods, PMSMT-Net achieves an average improvement of 3.7% in Dice coefficient and 3.6% in classification accuracy on two public BUS datasets. These results demonstrate the proposed model can significantly advance BUS-based tumor analysis and is of great significance for improving diagnostic precision and patient outcomes.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"110 ","pages":"Article 108212"},"PeriodicalIF":4.9000,"publicationDate":"2025-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-task learning for multi-scale breast cancer ultrasound image segmentation and classification based on visual perception\",\"authors\":\"Ao Su , Xiaolin Wang , Hao Xu , Jianfeng Zhang , Kang Chen , Dexing Kong , Guangfei Li , Xiaojun Chen , Jianming Wen , Zhong Lv\",\"doi\":\"10.1016/j.bspc.2025.108212\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Breast tumor segmentation and classification are essential components of breast ultrasound (BUS) computer-aided diagnosis (CAD) systems, which help improve the accuracy of breast cancer diagnoses. However, challenges arise due to the complexity of tumor features, the intensity similarities between lesions and surrounding tissues, as well as variations in tumor shape and location. While deep learning has been widely applied in CAD systems, many existing approaches overlook the relationship between segmentation and classification tasks, limiting their effectiveness. We propose PMSMT-Net, a Perception-based Multi-scale Ultrasound Image Segmentation and Classification Multi-task Learning Network, which enhances both segmentation and classification performance in BUS images. The segmentation network integrates a Visual Perception Module (VPM) to simulate human-like focus on regions of interest and, combined with Multi-scale Dilated Convolution (MSDC), accurately captures morphological, locational, and edge features. To further improve segmentation accuracy, we introduce the Variable Residual Convolutional Block Attention Module (VR-CBAM) and the Receptive Field Block-based Perceptually Separable Convolution Module (RFB-PSC), which enhance context feature fusion and reduce spatial information loss. The VPM output, along with the segmentation results, is then used as input to the classification network, where a transfer learning and ensemble learning approach classifies breast tumors. Compared to state-of-the-art methods, PMSMT-Net achieves an average improvement of 3.7% in Dice coefficient and 3.6% in classification accuracy on two public BUS datasets. These results demonstrate the proposed model can significantly advance BUS-based tumor analysis and is of great significance for improving diagnostic precision and patient outcomes.</div></div>\",\"PeriodicalId\":55362,\"journal\":{\"name\":\"Biomedical Signal Processing and Control\",\"volume\":\"110 \",\"pages\":\"Article 108212\"},\"PeriodicalIF\":4.9000,\"publicationDate\":\"2025-06-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Biomedical Signal Processing and Control\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1746809425007232\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, BIOMEDICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biomedical Signal Processing and Control","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1746809425007232","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
Multi-task learning for multi-scale breast cancer ultrasound image segmentation and classification based on visual perception
Breast tumor segmentation and classification are essential components of breast ultrasound (BUS) computer-aided diagnosis (CAD) systems, which help improve the accuracy of breast cancer diagnoses. However, challenges arise due to the complexity of tumor features, the intensity similarities between lesions and surrounding tissues, as well as variations in tumor shape and location. While deep learning has been widely applied in CAD systems, many existing approaches overlook the relationship between segmentation and classification tasks, limiting their effectiveness. We propose PMSMT-Net, a Perception-based Multi-scale Ultrasound Image Segmentation and Classification Multi-task Learning Network, which enhances both segmentation and classification performance in BUS images. The segmentation network integrates a Visual Perception Module (VPM) to simulate human-like focus on regions of interest and, combined with Multi-scale Dilated Convolution (MSDC), accurately captures morphological, locational, and edge features. To further improve segmentation accuracy, we introduce the Variable Residual Convolutional Block Attention Module (VR-CBAM) and the Receptive Field Block-based Perceptually Separable Convolution Module (RFB-PSC), which enhance context feature fusion and reduce spatial information loss. The VPM output, along with the segmentation results, is then used as input to the classification network, where a transfer learning and ensemble learning approach classifies breast tumors. Compared to state-of-the-art methods, PMSMT-Net achieves an average improvement of 3.7% in Dice coefficient and 3.6% in classification accuracy on two public BUS datasets. These results demonstrate the proposed model can significantly advance BUS-based tumor analysis and is of great significance for improving diagnostic precision and patient outcomes.
期刊介绍:
Biomedical Signal Processing and Control aims to provide a cross-disciplinary international forum for the interchange of information on research in the measurement and analysis of signals and images in clinical medicine and the biological sciences. Emphasis is placed on contributions dealing with the practical, applications-led research on the use of methods and devices in clinical diagnosis, patient monitoring and management.
Biomedical Signal Processing and Control reflects the main areas in which these methods are being used and developed at the interface of both engineering and clinical science. The scope of the journal is defined to include relevant review papers, technical notes, short communications and letters. Tutorial papers and special issues will also be published.