Yuzeng Chen , Qiangqiang Yuan , Yuqi Tang , Xin Wang , Yi Xiao , Jiang He , Ziyang Lihe , Xianyu Jin
{"title":"ProFiT: A prompt-guided frequency-aware filtering and template-enhanced interaction framework for hyperspectral video tracking","authors":"Yuzeng Chen , Qiangqiang Yuan , Yuqi Tang , Xin Wang , Yi Xiao , Jiang He , Ziyang Lihe , Xianyu Jin","doi":"10.1016/j.isprsjprs.2025.05.008","DOIUrl":null,"url":null,"abstract":"<div><div>Hyperspectral (HSP) video data can offer rich spectral-spatial–temporal information crucial for capturing object dynamics, attenuating the drawbacks of classical unimodal and multi-modal tracking. Current HSP tracking arts often suffer from feature refinements and information interactions, sealing the ceiling of capabilities. This study presents ProFiT, an innovative prompt-guided frequency-aware filtering and template-enhanced interaction framework for HSP video tracking, mitigating the above issues. First, ProFiT introduces a frequency-aware filtering module with adaptive filter generators to refine spectral-spatial consistency within HSP and false-color features. Then, a template-enhanced interaction module is introduced to extract complementary information for efficient cross-modal interactions. Furthermore, a token fusion module is devised to capture contextual dependencies with minimal parameters. While a temporal decoder embeds historical states, guiding to ensure temporal coherence. Comprehensive experiments across nine HSP benchmarks demonstrate that ProFiT achieves competitive accuracy, with overall precision and success rate scores of 0.870 and 0.678, respectively, along with a frame per second of 39.5. These results outperform 59 state-of-the-art trackers, establishing ProFiT as a robust solution for HSP video tracking. The code and result will be accessible at: <span><span>https://github.com/YZCU/ProFiT</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"226 ","pages":"Pages 164-186"},"PeriodicalIF":10.6000,"publicationDate":"2025-05-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271625001893","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Hyperspectral (HSP) video data can offer rich spectral-spatial–temporal information crucial for capturing object dynamics, attenuating the drawbacks of classical unimodal and multi-modal tracking. Current HSP tracking arts often suffer from feature refinements and information interactions, sealing the ceiling of capabilities. This study presents ProFiT, an innovative prompt-guided frequency-aware filtering and template-enhanced interaction framework for HSP video tracking, mitigating the above issues. First, ProFiT introduces a frequency-aware filtering module with adaptive filter generators to refine spectral-spatial consistency within HSP and false-color features. Then, a template-enhanced interaction module is introduced to extract complementary information for efficient cross-modal interactions. Furthermore, a token fusion module is devised to capture contextual dependencies with minimal parameters. While a temporal decoder embeds historical states, guiding to ensure temporal coherence. Comprehensive experiments across nine HSP benchmarks demonstrate that ProFiT achieves competitive accuracy, with overall precision and success rate scores of 0.870 and 0.678, respectively, along with a frame per second of 39.5. These results outperform 59 state-of-the-art trackers, establishing ProFiT as a robust solution for HSP video tracking. The code and result will be accessible at: https://github.com/YZCU/ProFiT.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.