Jinlong Xie , Liping Zhang , Long Cheng , Jian Yao , Pengjiang Qian , Binrong Zhu , Guiran Liu
{"title":"MARNet: Multi-scale adaptive relational network for robust point cloud completion via cross-modal fusion","authors":"Jinlong Xie , Liping Zhang , Long Cheng , Jian Yao , Pengjiang Qian , Binrong Zhu , Guiran Liu","doi":"10.1016/j.inffus.2025.103505","DOIUrl":null,"url":null,"abstract":"<div><div>Point cloud completion, pivotal for enabling robust 3D understanding in autonomous systems and augmented reality, faces persistent challenges in structural fidelity preservation and detail adaptive restoration. This paper presents MARNet—a novel multi-scale adaptive relational network via cross-modal guidance. We first design a spatial-relational adaptive feature descriptor integrating space-feature adaptive downsampling with relation-aware weighted edge convolution. This integration effectively preserves structural integrity while suppressing outlier interference. We then introduce a hierarchical cross-modal fusion module establishing bidirectional feature interaction pathways between 3D point clouds and multi-view images through attention-guided mechanisms, which significantly enhances feature representation capacity. Additionally, our adaptive multi-resolution point generator dynamically adjusts upsampling stages based on each shape’s geometric complexity, restoring highly detailed structures while mitigating over- and under-completion issues. Extensive experiments demonstrate state-of-the-art performance with Chamfer Distance of <span><math><mrow><mn>6</mn><mo>.</mo><mn>36</mn><mo>×</mo><mn>1</mn><msup><mrow><mn>0</mn></mrow><mrow><mn>4</mn></mrow></msup></mrow></math></span> on Completion 3D and <span><math><mrow><mn>6</mn><mo>.</mo><mn>42</mn><mo>×</mo><mn>1</mn><msup><mrow><mn>0</mn></mrow><mrow><mn>3</mn></mrow></msup></mrow></math></span> on PCN datasets. Our method outperforms existing approaches in preserving fine details and global consistency, particularly for complex structures, while exhibiting robustness to noise and viewpoint variations. The code is available at <span><span>https://github.com/long-git22/MARNet.git</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"126 ","pages":"Article 103505"},"PeriodicalIF":15.5000,"publicationDate":"2025-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525005779","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Point cloud completion, pivotal for enabling robust 3D understanding in autonomous systems and augmented reality, faces persistent challenges in structural fidelity preservation and detail adaptive restoration. This paper presents MARNet—a novel multi-scale adaptive relational network via cross-modal guidance. We first design a spatial-relational adaptive feature descriptor integrating space-feature adaptive downsampling with relation-aware weighted edge convolution. This integration effectively preserves structural integrity while suppressing outlier interference. We then introduce a hierarchical cross-modal fusion module establishing bidirectional feature interaction pathways between 3D point clouds and multi-view images through attention-guided mechanisms, which significantly enhances feature representation capacity. Additionally, our adaptive multi-resolution point generator dynamically adjusts upsampling stages based on each shape’s geometric complexity, restoring highly detailed structures while mitigating over- and under-completion issues. Extensive experiments demonstrate state-of-the-art performance with Chamfer Distance of on Completion 3D and on PCN datasets. Our method outperforms existing approaches in preserving fine details and global consistency, particularly for complex structures, while exhibiting robustness to noise and viewpoint variations. The code is available at https://github.com/long-git22/MARNet.git.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.