{"title":"DREB-Net: Dual-Stream Restoration Embedding Blur-Feature Fusion Network for High-Mobility UAV Object Detection","authors":"Qingpeng Li;Yuxin Zhang;Leyuan Fang;Yuhan Kang;Shutao Li;Xiao Xiang Zhu","doi":"10.1109/TGRS.2025.3543270","DOIUrl":null,"url":null,"abstract":"Object detection algorithms are pivotal components of UAV imaging systems, extensively employed in complex fields. However, images captured by high-mobility UAVs often suffer from motion blur cases, which significantly impedes the performance of advanced object detection algorithms. To address these challenges, we propose an innovative object detection algorithm specifically designed for blurry images, named dual-stream restoration embedding blur-feature fusion network (DREB-Net). First, DREB-Net addresses the particularities of blurry image object detection problem by incorporating a blurry image restoration auxiliary branch (BRAB) during the training phase. Second, it fuses the extracted shallow features via multilevel attention-guided feature fusion (MAGFF) module, to extract richer features. Here, the MAGFF module comprises local attention modules and global attention modules, which assign different weights to the branches. Then, during the inference phase, the deep feature extraction of the BRAB can be removed to reduce computational complexity and improve detection speed. In loss function, a combined loss of mean squared error (MSE) and SSIM is added to the BRAB to restore blurry images. Finally, DREB-Net introduces fast Fourier transform in the early stages of feature extraction, via a learnable frequency domain amplitude modulation module (LFAMM), to adjust feature amplitude and enhance feature processing capability. Compared to the baseline, DREB-Net achieved an approximate 7% increase in both mAP50 and mAR50 across two experimental datasets. Experimental results indicate that DREB-Net can still effectively perform object detection tasks under motion blur in captured images, showcasing excellent performance and broad application prospects. Our source code will be available at <uri>https://github.com/EEIC-Lab/DREB-Net.git</uri>.","PeriodicalId":13213,"journal":{"name":"IEEE Transactions on Geoscience and Remote Sensing","volume":"63 ","pages":"1-18"},"PeriodicalIF":7.5000,"publicationDate":"2025-02-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Geoscience and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10891913/","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Object detection algorithms are pivotal components of UAV imaging systems, extensively employed in complex fields. However, images captured by high-mobility UAVs often suffer from motion blur cases, which significantly impedes the performance of advanced object detection algorithms. To address these challenges, we propose an innovative object detection algorithm specifically designed for blurry images, named dual-stream restoration embedding blur-feature fusion network (DREB-Net). First, DREB-Net addresses the particularities of blurry image object detection problem by incorporating a blurry image restoration auxiliary branch (BRAB) during the training phase. Second, it fuses the extracted shallow features via multilevel attention-guided feature fusion (MAGFF) module, to extract richer features. Here, the MAGFF module comprises local attention modules and global attention modules, which assign different weights to the branches. Then, during the inference phase, the deep feature extraction of the BRAB can be removed to reduce computational complexity and improve detection speed. In loss function, a combined loss of mean squared error (MSE) and SSIM is added to the BRAB to restore blurry images. Finally, DREB-Net introduces fast Fourier transform in the early stages of feature extraction, via a learnable frequency domain amplitude modulation module (LFAMM), to adjust feature amplitude and enhance feature processing capability. Compared to the baseline, DREB-Net achieved an approximate 7% increase in both mAP50 and mAR50 across two experimental datasets. Experimental results indicate that DREB-Net can still effectively perform object detection tasks under motion blur in captured images, showcasing excellent performance and broad application prospects. Our source code will be available at https://github.com/EEIC-Lab/DREB-Net.git.
期刊介绍:
IEEE Transactions on Geoscience and Remote Sensing (TGRS) is a monthly publication that focuses on the theory, concepts, and techniques of science and engineering as applied to sensing the land, oceans, atmosphere, and space; and the processing, interpretation, and dissemination of this information.