{"title":"Towards Effective Rotation Generalization in UAV Object Re-Identification","authors":"Shuoyi Chen;Mang Ye;Yan Huang;Bo Du","doi":"10.1109/TIFS.2025.3544088","DOIUrl":null,"url":null,"abstract":"UAV surveillance offers a unique aerial perspective, enabling the monitoring of large areas and capturing targets from angles that fixed ground cameras cannot achieve. UAV-based object re-identification (ReID) differs from the extensively studied city camera scenarios, as it involves identifying specific objects in aerial images captured from a dynamic bird’s-eye view. The challenge lies in the significant variation in object perspectives and the often uncertain rotational changes captured by UAVs. Existing ReID methods designed for city cameras struggle to adapt to these rotational variations. To address these challenges, we propose a Transformer-based learnable rotation generalization enhancement method specifically for UAV-based ReID. To improve the model’s adaptability to uncertain rotational changes, we introduce a learnable feature-level rotation simulation technique that generates multiple rotated features. Building on this, we design a rotation diversification loss to decorrelate different rotated features, ensuring a rich feature representation. Additionally, to mitigate the negative effects of image-level rotation augmentation, we propose instance-level and distribution-level rotation invariance regularization. This approach establishes explicit associations between images and their rotated counterparts, facilitating the learning of visually consistent rotation-invariant features. Instance-level constraints ensure that detailed features remain consistent during rotation, while distribution-level constraints maintain the model’s semantic understanding. Notably, our method demonstrates strong versatility, covering a wide range of objects, including persons, vehicles, and various animals. Evaluations on multiple UAV-collected person and vehicle ReID datasets, as well as several animal datasets, consistently show outstanding performance, underscoring its robustness and adaptability to the unique challenges posed by UAV-based ReID.","PeriodicalId":13492,"journal":{"name":"IEEE Transactions on Information Forensics and Security","volume":"20 ","pages":"2593-2606"},"PeriodicalIF":6.3000,"publicationDate":"2025-02-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Information Forensics and Security","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10896735/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0
Abstract
UAV surveillance offers a unique aerial perspective, enabling the monitoring of large areas and capturing targets from angles that fixed ground cameras cannot achieve. UAV-based object re-identification (ReID) differs from the extensively studied city camera scenarios, as it involves identifying specific objects in aerial images captured from a dynamic bird’s-eye view. The challenge lies in the significant variation in object perspectives and the often uncertain rotational changes captured by UAVs. Existing ReID methods designed for city cameras struggle to adapt to these rotational variations. To address these challenges, we propose a Transformer-based learnable rotation generalization enhancement method specifically for UAV-based ReID. To improve the model’s adaptability to uncertain rotational changes, we introduce a learnable feature-level rotation simulation technique that generates multiple rotated features. Building on this, we design a rotation diversification loss to decorrelate different rotated features, ensuring a rich feature representation. Additionally, to mitigate the negative effects of image-level rotation augmentation, we propose instance-level and distribution-level rotation invariance regularization. This approach establishes explicit associations between images and their rotated counterparts, facilitating the learning of visually consistent rotation-invariant features. Instance-level constraints ensure that detailed features remain consistent during rotation, while distribution-level constraints maintain the model’s semantic understanding. Notably, our method demonstrates strong versatility, covering a wide range of objects, including persons, vehicles, and various animals. Evaluations on multiple UAV-collected person and vehicle ReID datasets, as well as several animal datasets, consistently show outstanding performance, underscoring its robustness and adaptability to the unique challenges posed by UAV-based ReID.
期刊介绍:
The IEEE Transactions on Information Forensics and Security covers the sciences, technologies, and applications relating to information forensics, information security, biometrics, surveillance and systems applications that incorporate these features