Jianhao Xu , Xiangtao Fan , Hongdeng Jian , Chen Xu , Weijia Bei , Qifeng Ge , Teng Zhao , Ruijie Han
{"title":"TAM-TR: Text-guided attention multi-modal transformer for object detection in UAV images","authors":"Jianhao Xu , Xiangtao Fan , Hongdeng Jian , Chen Xu , Weijia Bei , Qifeng Ge , Teng Zhao , Ruijie Han","doi":"10.1016/j.isprsjprs.2025.04.027","DOIUrl":null,"url":null,"abstract":"<div><div>Object detection in unmanned aerial vehicles (UAV) imagery is crucial in many fields, such as maritime search and rescue, remote sensing mapping, urban management and agricultural monitoring. The diverse perspectives and altitudes of UAV images often result in significant variations in the appearance and dimensions of objects, and occlusions are found more frequently than in general scenes. The unique bird’s-eye view of drones makes it more difficult for existing object detection models to distinguish between similar objects. A text-guided attention multi-modal transformer network named TAM-TR is proposed to address the above challenges. A Bidirectional Text–Image Attention Path Aggregation Network (BTA-PAN) is proposed in TAM-TR. This network imitates the architecture of the classic algorithm Scale-Invariant Feature Transform (SIFT) and shows better scale adaptability. A novel Multi-modal encoder–decoder head (MEH) was proposed, which can simultaneously consider all input sequence positions to avoid the disappearance of features of occluded objects. An additional text-guided attention branch, combined with a large text model, was proposed to improve the TAM-TR’s classification accuracy. Additionally, a Rotation-invariant IOU (RIOU) loss function was proposed to eliminate the previous loss function’s rotational instability. The experiment demonstrated that the TAM-TR outperformed the baseline by 9.5% and achieves 39.7% mean Averaged Precision (mAP) on the Visdrone dataset. The code will be available at <span><span>https://github.com/Xjh-UCAS/TAM-TR</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"227 ","pages":"Pages 170-184"},"PeriodicalIF":10.6000,"publicationDate":"2025-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271625001637","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Object detection in unmanned aerial vehicles (UAV) imagery is crucial in many fields, such as maritime search and rescue, remote sensing mapping, urban management and agricultural monitoring. The diverse perspectives and altitudes of UAV images often result in significant variations in the appearance and dimensions of objects, and occlusions are found more frequently than in general scenes. The unique bird’s-eye view of drones makes it more difficult for existing object detection models to distinguish between similar objects. A text-guided attention multi-modal transformer network named TAM-TR is proposed to address the above challenges. A Bidirectional Text–Image Attention Path Aggregation Network (BTA-PAN) is proposed in TAM-TR. This network imitates the architecture of the classic algorithm Scale-Invariant Feature Transform (SIFT) and shows better scale adaptability. A novel Multi-modal encoder–decoder head (MEH) was proposed, which can simultaneously consider all input sequence positions to avoid the disappearance of features of occluded objects. An additional text-guided attention branch, combined with a large text model, was proposed to improve the TAM-TR’s classification accuracy. Additionally, a Rotation-invariant IOU (RIOU) loss function was proposed to eliminate the previous loss function’s rotational instability. The experiment demonstrated that the TAM-TR outperformed the baseline by 9.5% and achieves 39.7% mean Averaged Precision (mAP) on the Visdrone dataset. The code will be available at https://github.com/Xjh-UCAS/TAM-TR.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.