{"title":"Adaptive and Background-Aware Match for Class-Agnostic Counting","authors":"Shenjian Gong;Jian Yang;Shanshan Zhang","doi":"10.1109/LSP.2025.3546891","DOIUrl":null,"url":null,"abstract":"Class-Agnostic Counting (CAC) aims to count object instances in an image by simply specifying a few exemplar boxes of interest. The key challenge for CAC is how to tailor a desirable interaction between exemplar and query features. Previous CAC methods implement such interaction by solely leveraging standard global feature convolution. We find this interaction leads to under-match caused by intra-class diversity and over-match on background, which harms counting performance severely. In this work, we propose a novel feature interaction method called Adaptive and Background-aware Match (ABM) against high intra-class diversity and noisy background. Concretely, given exemplar and query features, we improve the original high-dimensional coupled spaces match to Adaptive Orthogonal subspaces Match (AOM), avoiding under-match caused by intra-class diversity. Moreover, Background-Specific Match (BSM) employs interaction between the learnable background prototype and query features to provide global background priors, making the match be aware of background regions. Additionally, we find the high scale variance among different query images leads to bad counting performance for extremely small scale objects. Object-Scale Unify (OSU) is proposed to take the size of the exemplars as the scale prior and resize query images so that all objects are at a uniform average scale. Extensive experiments on FSC-147 show that our method performs better. We also conduct extensive ablation studies to demonstrate the effectiveness of each component of our proposed method.","PeriodicalId":13154,"journal":{"name":"IEEE Signal Processing Letters","volume":"32 ","pages":"1261-1265"},"PeriodicalIF":3.2000,"publicationDate":"2025-03-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Signal Processing Letters","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10908813/","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Class-Agnostic Counting (CAC) aims to count object instances in an image by simply specifying a few exemplar boxes of interest. The key challenge for CAC is how to tailor a desirable interaction between exemplar and query features. Previous CAC methods implement such interaction by solely leveraging standard global feature convolution. We find this interaction leads to under-match caused by intra-class diversity and over-match on background, which harms counting performance severely. In this work, we propose a novel feature interaction method called Adaptive and Background-aware Match (ABM) against high intra-class diversity and noisy background. Concretely, given exemplar and query features, we improve the original high-dimensional coupled spaces match to Adaptive Orthogonal subspaces Match (AOM), avoiding under-match caused by intra-class diversity. Moreover, Background-Specific Match (BSM) employs interaction between the learnable background prototype and query features to provide global background priors, making the match be aware of background regions. Additionally, we find the high scale variance among different query images leads to bad counting performance for extremely small scale objects. Object-Scale Unify (OSU) is proposed to take the size of the exemplars as the scale prior and resize query images so that all objects are at a uniform average scale. Extensive experiments on FSC-147 show that our method performs better. We also conduct extensive ablation studies to demonstrate the effectiveness of each component of our proposed method.
期刊介绍:
The IEEE Signal Processing Letters is a monthly, archival publication designed to provide rapid dissemination of original, cutting-edge ideas and timely, significant contributions in signal, image, speech, language and audio processing. Papers published in the Letters can be presented within one year of their appearance in signal processing conferences such as ICASSP, GlobalSIP and ICIP, and also in several workshop organized by the Signal Processing Society.