{"title":"H-vmunet: High-order Vision Mamba UNet for medical image segmentation","authors":"Renkai Wu , Yinghao Liu , Pengchen Liang , Qing Chang","doi":"10.1016/j.neucom.2025.129447","DOIUrl":null,"url":null,"abstract":"<div><div>In the field of medical image segmentation, variant models based on Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs) have been extensively developed and applied. However, CNNs often struggle with processing long-sequence information, while Vision Transformers exhibit low sensitivity to local feature extraction and face challenges with computational complexity. Recently, the emergence of state-space models (SSMs), particularly 2D-selective-scan (SS2D), has challenged the longtime dominance of traditional CNNs and ViTs as foundational modules in visual neural networks. In this paper, we extend the adaptation of SS2D by proposing a High-order Vision Mamba UNet (H-vmunet) model for medical image segmentation. Among them, the H-vmunet model includes the proposed novel High-order 2D-selective-scan (H-SS2D) and Local-SS2D module. The H-SS2D is used to reduce the introduction of redundant information when SS2D features are feature-learning in the global receptive field. The Local-SS2D module is used to improve the learning ability of local features in SS2D. We conducted comprehensive comparison and ablation experiments on three publicly available medical image datasets (ISIC2017, Spleen, and CVC-ClinicDB), and the results consistently demonstrate the strong performance of H-vmunet in medical image segmentation tasks. The code is available from <span><span>https://github.com/wurenkai/H-vmunet</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"624 ","pages":"Article 129447"},"PeriodicalIF":5.5000,"publicationDate":"2025-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225001195","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
In the field of medical image segmentation, variant models based on Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs) have been extensively developed and applied. However, CNNs often struggle with processing long-sequence information, while Vision Transformers exhibit low sensitivity to local feature extraction and face challenges with computational complexity. Recently, the emergence of state-space models (SSMs), particularly 2D-selective-scan (SS2D), has challenged the longtime dominance of traditional CNNs and ViTs as foundational modules in visual neural networks. In this paper, we extend the adaptation of SS2D by proposing a High-order Vision Mamba UNet (H-vmunet) model for medical image segmentation. Among them, the H-vmunet model includes the proposed novel High-order 2D-selective-scan (H-SS2D) and Local-SS2D module. The H-SS2D is used to reduce the introduction of redundant information when SS2D features are feature-learning in the global receptive field. The Local-SS2D module is used to improve the learning ability of local features in SS2D. We conducted comprehensive comparison and ablation experiments on three publicly available medical image datasets (ISIC2017, Spleen, and CVC-ClinicDB), and the results consistently demonstrate the strong performance of H-vmunet in medical image segmentation tasks. The code is available from https://github.com/wurenkai/H-vmunet.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.