Wenguang Tao;Xiaotian Wang;Tian Yan;Haixia Bi;Jie Yan
{"title":"遥感图像中微小目标的编码器-解码器域增强对准检测器","authors":"Wenguang Tao;Xiaotian Wang;Tian Yan;Haixia Bi;Jie Yan","doi":"10.1109/TGRS.2024.3510948","DOIUrl":null,"url":null,"abstract":"In recent years, deep learning has shown great potential in object detection applications, but it is still difficult to accurately detect tiny objects with an area proportion of less than 1% in remote sensing images. Most existing studies focus on designing complex networks to learn discriminative features of tiny objects, usually resulting in a heavy computational burden. In contrast, this article proposes an accurate and efficient single-stage detector called EDADet for tiny objects. First, domain conversion technology is used to realize cross-domain multimodal data fusion based on single-modal data input. Then, a tiny object-aware backbone is designed to extract features at different scales. Next, an encoder–decoder feature fusion (EDFF) structure is devised to achieve efficient cross-scale propagation of semantic information. Finally, a center-assist loss and an alignment self-supervised loss are adopted to alleviate the position sensitivity issue and drift of tiny objects. A series of experiments on the AI-TODv2 dataset demonstrate the effectiveness and practicality of our EDADet. It achieves state-of-the-art (SOTA) performance and surpasses the second-best method by 9.65% in AP50 and 4.86% in mAP.","PeriodicalId":13213,"journal":{"name":"IEEE Transactions on Geoscience and Remote Sensing","volume":"63 ","pages":"1-15"},"PeriodicalIF":8.6000,"publicationDate":"2024-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"EDADet: Encoder–Decoder Domain Augmented Alignment Detector for Tiny Objects in Remote Sensing Images\",\"authors\":\"Wenguang Tao;Xiaotian Wang;Tian Yan;Haixia Bi;Jie Yan\",\"doi\":\"10.1109/TGRS.2024.3510948\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In recent years, deep learning has shown great potential in object detection applications, but it is still difficult to accurately detect tiny objects with an area proportion of less than 1% in remote sensing images. Most existing studies focus on designing complex networks to learn discriminative features of tiny objects, usually resulting in a heavy computational burden. In contrast, this article proposes an accurate and efficient single-stage detector called EDADet for tiny objects. First, domain conversion technology is used to realize cross-domain multimodal data fusion based on single-modal data input. Then, a tiny object-aware backbone is designed to extract features at different scales. Next, an encoder–decoder feature fusion (EDFF) structure is devised to achieve efficient cross-scale propagation of semantic information. Finally, a center-assist loss and an alignment self-supervised loss are adopted to alleviate the position sensitivity issue and drift of tiny objects. A series of experiments on the AI-TODv2 dataset demonstrate the effectiveness and practicality of our EDADet. It achieves state-of-the-art (SOTA) performance and surpasses the second-best method by 9.65% in AP50 and 4.86% in mAP.\",\"PeriodicalId\":13213,\"journal\":{\"name\":\"IEEE Transactions on Geoscience and Remote Sensing\",\"volume\":\"63 \",\"pages\":\"1-15\"},\"PeriodicalIF\":8.6000,\"publicationDate\":\"2024-12-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Geoscience and Remote Sensing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10777552/\",\"RegionNum\":1,\"RegionCategory\":\"地球科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Geoscience and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10777552/","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
EDADet: Encoder–Decoder Domain Augmented Alignment Detector for Tiny Objects in Remote Sensing Images
In recent years, deep learning has shown great potential in object detection applications, but it is still difficult to accurately detect tiny objects with an area proportion of less than 1% in remote sensing images. Most existing studies focus on designing complex networks to learn discriminative features of tiny objects, usually resulting in a heavy computational burden. In contrast, this article proposes an accurate and efficient single-stage detector called EDADet for tiny objects. First, domain conversion technology is used to realize cross-domain multimodal data fusion based on single-modal data input. Then, a tiny object-aware backbone is designed to extract features at different scales. Next, an encoder–decoder feature fusion (EDFF) structure is devised to achieve efficient cross-scale propagation of semantic information. Finally, a center-assist loss and an alignment self-supervised loss are adopted to alleviate the position sensitivity issue and drift of tiny objects. A series of experiments on the AI-TODv2 dataset demonstrate the effectiveness and practicality of our EDADet. It achieves state-of-the-art (SOTA) performance and surpasses the second-best method by 9.65% in AP50 and 4.86% in mAP.
期刊介绍:
IEEE Transactions on Geoscience and Remote Sensing (TGRS) is a monthly publication that focuses on the theory, concepts, and techniques of science and engineering as applied to sensing the land, oceans, atmosphere, and space; and the processing, interpretation, and dissemination of this information.