Ju Zhong;Aimin Jiang;Chang Liu;Ning Xu;Yanping Zhu
{"title":"Depth Completion With Super-Resolution and Cross-Modality Optimization","authors":"Ju Zhong;Aimin Jiang;Chang Liu;Ning Xu;Yanping Zhu","doi":"10.1109/LRA.2025.3560860","DOIUrl":null,"url":null,"abstract":"Depth completion, the process of generating dense depth maps from sparse or incomplete data, is inherently challenging due to the variability in sensor types, environmental conditions, and the differences between data modalities. In this letter, we propose a novel framework for depth completion. First, we introduce an efficient depth estimation network capable of predicting relative depth from a single RGB image. Next, we design a depth super-resolution network that refines the predicted depth by using Fast Fourier Convolution (FFC) and Gradient-weighted Symmetric Feature Transmission (GSFT) modules. These modules upsample the depth map using high-resolution RGB guidance, effectively mitigating the cross-modality gap. Finally, a global optimization step fuses the upsampled depth with sparse ground truth to produce high-quality dense depth maps. Our unified approach enhances generalization across diverse datasets while avoiding overfitting to specific depth corruption patterns. The enhancement in depth resolution and accuracy are critical for robotic applications requiring precise spatial perception, such as localization and manipulation. Experimental results on NYU-Depth V2 and SUN RGB-D benchmarks demonstrate the superiority of our approach compared to state-of-the-art approaches.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"10 6","pages":"5585-5592"},"PeriodicalIF":4.6000,"publicationDate":"2025-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Robotics and Automation Letters","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10965466/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0
Abstract
Depth completion, the process of generating dense depth maps from sparse or incomplete data, is inherently challenging due to the variability in sensor types, environmental conditions, and the differences between data modalities. In this letter, we propose a novel framework for depth completion. First, we introduce an efficient depth estimation network capable of predicting relative depth from a single RGB image. Next, we design a depth super-resolution network that refines the predicted depth by using Fast Fourier Convolution (FFC) and Gradient-weighted Symmetric Feature Transmission (GSFT) modules. These modules upsample the depth map using high-resolution RGB guidance, effectively mitigating the cross-modality gap. Finally, a global optimization step fuses the upsampled depth with sparse ground truth to produce high-quality dense depth maps. Our unified approach enhances generalization across diverse datasets while avoiding overfitting to specific depth corruption patterns. The enhancement in depth resolution and accuracy are critical for robotic applications requiring precise spatial perception, such as localization and manipulation. Experimental results on NYU-Depth V2 and SUN RGB-D benchmarks demonstrate the superiority of our approach compared to state-of-the-art approaches.
期刊介绍:
The scope of this journal is to publish peer-reviewed articles that provide a timely and concise account of innovative research ideas and application results, reporting significant theoretical findings and application case studies in areas of robotics and automation.