Haijun Meng , Huimin Lu , Bozhi Ding , Qiangchang Wang
{"title":"SPADesc: Semantic and parallel attention with feature description","authors":"Haijun Meng , Huimin Lu , Bozhi Ding , Qiangchang Wang","doi":"10.1016/j.neucom.2025.129567","DOIUrl":null,"url":null,"abstract":"<div><div>Local feature detection and description are essential preliminary tasks in a multitude of computer vision applications. Despite the prowess of deep neural networks in feature extraction, they still grapple with challenges in capturing globally invariant and robust features, especially in dynamic scenes and areas with simplistic and repetitive geometric structures. This paper introduces a multi-scale feature fusion framework, SPADesc, which addresses these challenges by leveraging dynamic weighted fusion (DWF) and semantic priors. We integrate convolutional and self-attention mechanisms to bolster local feature detection and description in complex environments. Our approach employs a Parallel Convolution and Attention (PCA) module to generate descriptors that encompass both local and global scales. Additionally, a Semantic-Guided (SG) module is employed to produce class-aware global mask information, which implicitly guides the selection of keypoints and descriptors. By incorporating a Semantically Weighted (SW) loss function, we enhance the robustness and discriminative power of the descriptors. Extensive experimental results across various visual tasks demonstrate significant performance improvements, highlighting the superior adaptability and precision of our proposed model. The code is available at <span><span>https://github.com/Diffcc/SPADesc</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"625 ","pages":"Article 129567"},"PeriodicalIF":5.5000,"publicationDate":"2025-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225002395","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Local feature detection and description are essential preliminary tasks in a multitude of computer vision applications. Despite the prowess of deep neural networks in feature extraction, they still grapple with challenges in capturing globally invariant and robust features, especially in dynamic scenes and areas with simplistic and repetitive geometric structures. This paper introduces a multi-scale feature fusion framework, SPADesc, which addresses these challenges by leveraging dynamic weighted fusion (DWF) and semantic priors. We integrate convolutional and self-attention mechanisms to bolster local feature detection and description in complex environments. Our approach employs a Parallel Convolution and Attention (PCA) module to generate descriptors that encompass both local and global scales. Additionally, a Semantic-Guided (SG) module is employed to produce class-aware global mask information, which implicitly guides the selection of keypoints and descriptors. By incorporating a Semantically Weighted (SW) loss function, we enhance the robustness and discriminative power of the descriptors. Extensive experimental results across various visual tasks demonstrate significant performance improvements, highlighting the superior adaptability and precision of our proposed model. The code is available at https://github.com/Diffcc/SPADesc.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.