{"title":"拉普拉斯注意:一种即插即用的算法,不会增加视觉任务的模型复杂度","authors":"Xiaolei Chen, Yubing Lu, Runyu Wen","doi":"10.1049/cit2.12402","DOIUrl":null,"url":null,"abstract":"<p>Most prevailing attention mechanism modules in contemporary research are convolution-based modules, and while these modules contribute to enhancing the accuracy of deep learning networks in visual tasks, they concurrently augment the overall model complexity. To address the problem, this paper proposes a plug-and-play algorithm that does not increase the complexity of the model, Laplacian attention (LA). The LA algorithm first calculates the similarity distance between feature points in the feature space and feature channel and constructs the residual Laplacian matrix between feature points through the similarity distance and Gaussian kernel. This construction serves to segregate non-similar feature points while aggregating those with similarities. Ultimately, the LA algorithm allocates the outputs of the feature channel and the feature space adaptively to derive the final LA outputs. Crucially, the LA algorithm is confined to the forward computation process and does not involve backpropagation or any parameter learning. The LA algorithm undergoes comprehensive experimentation on three distinct datasets—namely Cifar-10, miniImageNet, and Pascal VOC 2012. The experimental results demonstrate that, compared with the advanced attention mechanism modules in recent years, such as SENet, CBAM, ECANet, coordinate attention, and triplet attention, the LA algorithm exhibits superior performance across image classification, object detection and semantic segmentation tasks.</p>","PeriodicalId":46211,"journal":{"name":"CAAI Transactions on Intelligence Technology","volume":"10 2","pages":"545-556"},"PeriodicalIF":8.4000,"publicationDate":"2024-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12402","citationCount":"0","resultStr":"{\"title\":\"Laplacian attention: A plug-and-play algorithm without increasing model complexity for vision tasks\",\"authors\":\"Xiaolei Chen, Yubing Lu, Runyu Wen\",\"doi\":\"10.1049/cit2.12402\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Most prevailing attention mechanism modules in contemporary research are convolution-based modules, and while these modules contribute to enhancing the accuracy of deep learning networks in visual tasks, they concurrently augment the overall model complexity. To address the problem, this paper proposes a plug-and-play algorithm that does not increase the complexity of the model, Laplacian attention (LA). The LA algorithm first calculates the similarity distance between feature points in the feature space and feature channel and constructs the residual Laplacian matrix between feature points through the similarity distance and Gaussian kernel. This construction serves to segregate non-similar feature points while aggregating those with similarities. Ultimately, the LA algorithm allocates the outputs of the feature channel and the feature space adaptively to derive the final LA outputs. Crucially, the LA algorithm is confined to the forward computation process and does not involve backpropagation or any parameter learning. The LA algorithm undergoes comprehensive experimentation on three distinct datasets—namely Cifar-10, miniImageNet, and Pascal VOC 2012. The experimental results demonstrate that, compared with the advanced attention mechanism modules in recent years, such as SENet, CBAM, ECANet, coordinate attention, and triplet attention, the LA algorithm exhibits superior performance across image classification, object detection and semantic segmentation tasks.</p>\",\"PeriodicalId\":46211,\"journal\":{\"name\":\"CAAI Transactions on Intelligence Technology\",\"volume\":\"10 2\",\"pages\":\"545-556\"},\"PeriodicalIF\":8.4000,\"publicationDate\":\"2024-12-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cit2.12402\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"CAAI Transactions on Intelligence Technology\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1049/cit2.12402\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"CAAI Transactions on Intelligence Technology","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1049/cit2.12402","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Laplacian attention: A plug-and-play algorithm without increasing model complexity for vision tasks
Most prevailing attention mechanism modules in contemporary research are convolution-based modules, and while these modules contribute to enhancing the accuracy of deep learning networks in visual tasks, they concurrently augment the overall model complexity. To address the problem, this paper proposes a plug-and-play algorithm that does not increase the complexity of the model, Laplacian attention (LA). The LA algorithm first calculates the similarity distance between feature points in the feature space and feature channel and constructs the residual Laplacian matrix between feature points through the similarity distance and Gaussian kernel. This construction serves to segregate non-similar feature points while aggregating those with similarities. Ultimately, the LA algorithm allocates the outputs of the feature channel and the feature space adaptively to derive the final LA outputs. Crucially, the LA algorithm is confined to the forward computation process and does not involve backpropagation or any parameter learning. The LA algorithm undergoes comprehensive experimentation on three distinct datasets—namely Cifar-10, miniImageNet, and Pascal VOC 2012. The experimental results demonstrate that, compared with the advanced attention mechanism modules in recent years, such as SENet, CBAM, ECANet, coordinate attention, and triplet attention, the LA algorithm exhibits superior performance across image classification, object detection and semantic segmentation tasks.
期刊介绍:
CAAI Transactions on Intelligence Technology is a leading venue for original research on the theoretical and experimental aspects of artificial intelligence technology. We are a fully open access journal co-published by the Institution of Engineering and Technology (IET) and the Chinese Association for Artificial Intelligence (CAAI) providing research which is openly accessible to read and share worldwide.