Huimin Guo;Yin Gu;Wu Du;Boyang Chen;Ming Cui;Teng Zhang;Deyu Sun;Wei Qian;He Ma
{"title":"MCHNet:一种用于CT图像中危险器官分割的高效交叉注意引导分层多尺度网络","authors":"Huimin Guo;Yin Gu;Wu Du;Boyang Chen;Ming Cui;Teng Zhang;Deyu Sun;Wei Qian;He Ma","doi":"10.1109/TRPMS.2024.3509390","DOIUrl":null,"url":null,"abstract":"In radiotherapy, precisely contoured organs at risk (OARs) near the target areas are essential for effective treatment planning. Manual delineation of OARs is labor-intensive and varies among experts. Deep learning has improved accuracy and consistency, but current methods use complex architectures with many parameters, risking overfitting on small medical image datasets. Addressing this, this article introduces MCHNet, a cross attention (CA)-guided hierarchical multiscale segmentation network based on CT images. MCHNet adopts a typical U-shaped structure, with MobileVit as its backbone. An efficient CA-guided block is utilized to enhance feature extraction while minimizing model parameters. Additionally, a novel skip-connection strategy is proposed to preserve critical medical image information during multiple down-sampling operations and bridge the gap between deep and shallow features. We conducted extensive experiments on three public datasets, i.e., multiatlas labeling beyond the cranial vault dataset, Segmentation of thoracic OARs dataset, and multimodality abdominal multiorgan segmentation challenge 2022 dataset. Experiment results demonstrate that the proposed method outperforms other related Transformer-based or hybrid models in terms of computational complexity, quantitative and qualitative results. We believe that the proposed method can offer an optimal blend of precision and efficiency that advances the capabilities of OARs segmentation in radiotherapy.","PeriodicalId":46807,"journal":{"name":"IEEE Transactions on Radiation and Plasma Medical Sciences","volume":"9 5","pages":"598-612"},"PeriodicalIF":4.6000,"publicationDate":"2024-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"MCHNet: An Efficient Cross Attention-Guided Hierarchical Multiscale Network for Segmentation of Organs at Risk in CT Images\",\"authors\":\"Huimin Guo;Yin Gu;Wu Du;Boyang Chen;Ming Cui;Teng Zhang;Deyu Sun;Wei Qian;He Ma\",\"doi\":\"10.1109/TRPMS.2024.3509390\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In radiotherapy, precisely contoured organs at risk (OARs) near the target areas are essential for effective treatment planning. Manual delineation of OARs is labor-intensive and varies among experts. Deep learning has improved accuracy and consistency, but current methods use complex architectures with many parameters, risking overfitting on small medical image datasets. Addressing this, this article introduces MCHNet, a cross attention (CA)-guided hierarchical multiscale segmentation network based on CT images. MCHNet adopts a typical U-shaped structure, with MobileVit as its backbone. An efficient CA-guided block is utilized to enhance feature extraction while minimizing model parameters. Additionally, a novel skip-connection strategy is proposed to preserve critical medical image information during multiple down-sampling operations and bridge the gap between deep and shallow features. We conducted extensive experiments on three public datasets, i.e., multiatlas labeling beyond the cranial vault dataset, Segmentation of thoracic OARs dataset, and multimodality abdominal multiorgan segmentation challenge 2022 dataset. Experiment results demonstrate that the proposed method outperforms other related Transformer-based or hybrid models in terms of computational complexity, quantitative and qualitative results. We believe that the proposed method can offer an optimal blend of precision and efficiency that advances the capabilities of OARs segmentation in radiotherapy.\",\"PeriodicalId\":46807,\"journal\":{\"name\":\"IEEE Transactions on Radiation and Plasma Medical Sciences\",\"volume\":\"9 5\",\"pages\":\"598-612\"},\"PeriodicalIF\":4.6000,\"publicationDate\":\"2024-12-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Radiation and Plasma Medical Sciences\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10772221/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Radiation and Plasma Medical Sciences","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10772221/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
MCHNet: An Efficient Cross Attention-Guided Hierarchical Multiscale Network for Segmentation of Organs at Risk in CT Images
In radiotherapy, precisely contoured organs at risk (OARs) near the target areas are essential for effective treatment planning. Manual delineation of OARs is labor-intensive and varies among experts. Deep learning has improved accuracy and consistency, but current methods use complex architectures with many parameters, risking overfitting on small medical image datasets. Addressing this, this article introduces MCHNet, a cross attention (CA)-guided hierarchical multiscale segmentation network based on CT images. MCHNet adopts a typical U-shaped structure, with MobileVit as its backbone. An efficient CA-guided block is utilized to enhance feature extraction while minimizing model parameters. Additionally, a novel skip-connection strategy is proposed to preserve critical medical image information during multiple down-sampling operations and bridge the gap between deep and shallow features. We conducted extensive experiments on three public datasets, i.e., multiatlas labeling beyond the cranial vault dataset, Segmentation of thoracic OARs dataset, and multimodality abdominal multiorgan segmentation challenge 2022 dataset. Experiment results demonstrate that the proposed method outperforms other related Transformer-based or hybrid models in terms of computational complexity, quantitative and qualitative results. We believe that the proposed method can offer an optimal blend of precision and efficiency that advances the capabilities of OARs segmentation in radiotherapy.