{"title":"Swim-Rep融合网络:一种具有更快循环交叉极化注意的新骨干。","authors":"Zhe Li","doi":"10.1371/journal.pone.0321270","DOIUrl":null,"url":null,"abstract":"<p><p>Deep learning techniques are widely used in the field of medicine and image classification. In past studies, SwimTransformer and RepVGG are very efficient and classical deep learning models. Multi-scale feature fusion and attention mechanisms are effective means to enhance the performance of deep learning models. In this paper, we introduce a novel Swim-Rep fusion network, along with a new multi-scale feature fusion module called multi-scale strip pooling fusion module(MPF) and a new attention module called Faster Recurrent Criss Cross Polarized Attention (FRCPA), both of which excel at extracting multi-dimensional cross-attention and fine-grained features. Our fully supervised model achieved an impressive accuracy of 99.82% on the MIT-BIH database, outperforming the ViT model classifier by 0.12%. Additionally, our semi-supervised model demonstrated strong performance, achieving 98.4% accuracy on the validation set. Experimental results on the remote sensing image classification dataset RSSCN7 demonstrate that our new base model achieves a classification accuracy of 92.5%, which is 8.57% better than the classification performance of swim-transformer-base and 12.9% better than that of RepVGG-base, and increasing the depth of the module yields superior performance.</p>","PeriodicalId":20189,"journal":{"name":"PLoS ONE","volume":"20 5","pages":"e0321270"},"PeriodicalIF":2.9000,"publicationDate":"2025-05-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Swim-Rep fusion net: A new backbone with Faster Recurrent Criss Cross Polarized Attention.\",\"authors\":\"Zhe Li\",\"doi\":\"10.1371/journal.pone.0321270\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Deep learning techniques are widely used in the field of medicine and image classification. In past studies, SwimTransformer and RepVGG are very efficient and classical deep learning models. Multi-scale feature fusion and attention mechanisms are effective means to enhance the performance of deep learning models. In this paper, we introduce a novel Swim-Rep fusion network, along with a new multi-scale feature fusion module called multi-scale strip pooling fusion module(MPF) and a new attention module called Faster Recurrent Criss Cross Polarized Attention (FRCPA), both of which excel at extracting multi-dimensional cross-attention and fine-grained features. Our fully supervised model achieved an impressive accuracy of 99.82% on the MIT-BIH database, outperforming the ViT model classifier by 0.12%. Additionally, our semi-supervised model demonstrated strong performance, achieving 98.4% accuracy on the validation set. Experimental results on the remote sensing image classification dataset RSSCN7 demonstrate that our new base model achieves a classification accuracy of 92.5%, which is 8.57% better than the classification performance of swim-transformer-base and 12.9% better than that of RepVGG-base, and increasing the depth of the module yields superior performance.</p>\",\"PeriodicalId\":20189,\"journal\":{\"name\":\"PLoS ONE\",\"volume\":\"20 5\",\"pages\":\"e0321270\"},\"PeriodicalIF\":2.9000,\"publicationDate\":\"2025-05-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"PLoS ONE\",\"FirstCategoryId\":\"103\",\"ListUrlMain\":\"https://doi.org/10.1371/journal.pone.0321270\",\"RegionNum\":3,\"RegionCategory\":\"综合性期刊\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/1/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"Q1\",\"JCRName\":\"MULTIDISCIPLINARY SCIENCES\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"PLoS ONE","FirstCategoryId":"103","ListUrlMain":"https://doi.org/10.1371/journal.pone.0321270","RegionNum":3,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/1 0:00:00","PubModel":"eCollection","JCR":"Q1","JCRName":"MULTIDISCIPLINARY SCIENCES","Score":null,"Total":0}
Swim-Rep fusion net: A new backbone with Faster Recurrent Criss Cross Polarized Attention.
Deep learning techniques are widely used in the field of medicine and image classification. In past studies, SwimTransformer and RepVGG are very efficient and classical deep learning models. Multi-scale feature fusion and attention mechanisms are effective means to enhance the performance of deep learning models. In this paper, we introduce a novel Swim-Rep fusion network, along with a new multi-scale feature fusion module called multi-scale strip pooling fusion module(MPF) and a new attention module called Faster Recurrent Criss Cross Polarized Attention (FRCPA), both of which excel at extracting multi-dimensional cross-attention and fine-grained features. Our fully supervised model achieved an impressive accuracy of 99.82% on the MIT-BIH database, outperforming the ViT model classifier by 0.12%. Additionally, our semi-supervised model demonstrated strong performance, achieving 98.4% accuracy on the validation set. Experimental results on the remote sensing image classification dataset RSSCN7 demonstrate that our new base model achieves a classification accuracy of 92.5%, which is 8.57% better than the classification performance of swim-transformer-base and 12.9% better than that of RepVGG-base, and increasing the depth of the module yields superior performance.
期刊介绍:
PLOS ONE is an international, peer-reviewed, open-access, online publication. PLOS ONE welcomes reports on primary research from any scientific discipline. It provides:
* Open-access—freely accessible online, authors retain copyright
* Fast publication times
* Peer review by expert, practicing researchers
* Post-publication tools to indicate quality and impact
* Community-based dialogue on articles
* Worldwide media coverage