{"title":"A fused attention-based hybrid model for semi-supervised medical image segmentation","authors":"Masum Shah Junayed, Sheida Nabavi","doi":"10.1016/j.bea.2025.100203","DOIUrl":null,"url":null,"abstract":"<div><div>Addressing the significant challenge of extensive data annotation in medical image segmentation, semi-supervised techniques have become an effective alternative for leveraging both labeled and unlabeled data. However, accurately capturing complex local structures and global contextual information remains difficult, often leading to inconsistent segmentation results. To mitigate these limitations, we propose a fused transformer-based hybrid architecture for semi-supervised medical image segmentation. The model integrates a parallel backbone comprising Deformation Convolution Blocks (DCB) and Fused Transformer Blocks (FTB), followed by a segmentation head for precise mask prediction. The DCB enhances spatial adaptability for irregular and artifact-rich regions, while the FTB – with its fused attention mechanism – captures long-range dependencies efficiently. A Ghost Layer Perceptron (GLP) embedded within the transformer further improves computational efficiency without compromising representation quality. In addition, the incorporation of a consistency loss and unsupervised contrastive learning facilitates robust feature discrimination on unlabeled data, improving generalization across modalities. Extensive experiments on four publicly available medical imaging datasets demonstrate that the proposed model achieves comparable or better accuracy than recent state-of-the-art methods, while requiring substantially fewer parameters and lower computational cost, underscoring its practicality for real-world clinical applications.</div></div>","PeriodicalId":72384,"journal":{"name":"Biomedical engineering advances","volume":"11 ","pages":"Article 100203"},"PeriodicalIF":0.0000,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biomedical engineering advances","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2667099225000593","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/12/22 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Addressing the significant challenge of extensive data annotation in medical image segmentation, semi-supervised techniques have become an effective alternative for leveraging both labeled and unlabeled data. However, accurately capturing complex local structures and global contextual information remains difficult, often leading to inconsistent segmentation results. To mitigate these limitations, we propose a fused transformer-based hybrid architecture for semi-supervised medical image segmentation. The model integrates a parallel backbone comprising Deformation Convolution Blocks (DCB) and Fused Transformer Blocks (FTB), followed by a segmentation head for precise mask prediction. The DCB enhances spatial adaptability for irregular and artifact-rich regions, while the FTB – with its fused attention mechanism – captures long-range dependencies efficiently. A Ghost Layer Perceptron (GLP) embedded within the transformer further improves computational efficiency without compromising representation quality. In addition, the incorporation of a consistency loss and unsupervised contrastive learning facilitates robust feature discrimination on unlabeled data, improving generalization across modalities. Extensive experiments on four publicly available medical imaging datasets demonstrate that the proposed model achieves comparable or better accuracy than recent state-of-the-art methods, while requiring substantially fewer parameters and lower computational cost, underscoring its practicality for real-world clinical applications.