Tingyu Zhang;Zhigang Liang;Yanzhao Yang;Xinyu Yang;Yu Zhu;Jian Wang
{"title":"三维目标检测的对比后期融合","authors":"Tingyu Zhang;Zhigang Liang;Yanzhao Yang;Xinyu Yang;Yu Zhu;Jian Wang","doi":"10.1109/TIV.2024.3454085","DOIUrl":null,"url":null,"abstract":"In the field of autonomous driving, accurate and efficient 3D object detection is crucial for ensuring safe and reliable operation. This paper focuses on the fusion of camera and LiDAR data in a late-fusion manner for 3D object detection. The proposed approach incorporates contrastive learning to enhance feature consistency between camera and LiDAR candidates, which is named as Contrastive Camera-LiDAR Object Candidates (C-CLOCs) fusion network, facilitating better fusion results. We delve into the label assignment aspect in late fusion methods and introduce a novel label assignment strategy to filter out irrelevant information. Additionally, a Multi-modality Ground-truth Sampling (MGS) method is introduced, which leverages the inclusion of point cloud information from LiDAR and corresponding images in training samples, resulting in improved performance. Experimental results demonstrate the effectiveness of the proposed method in achieving accurate 3D object detection in autonomous driving scenarios.","PeriodicalId":36532,"journal":{"name":"IEEE Transactions on Intelligent Vehicles","volume":"10 5","pages":"3442-3457"},"PeriodicalIF":14.3000,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Contrastive Late Fusion for 3D Object Detection\",\"authors\":\"Tingyu Zhang;Zhigang Liang;Yanzhao Yang;Xinyu Yang;Yu Zhu;Jian Wang\",\"doi\":\"10.1109/TIV.2024.3454085\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In the field of autonomous driving, accurate and efficient 3D object detection is crucial for ensuring safe and reliable operation. This paper focuses on the fusion of camera and LiDAR data in a late-fusion manner for 3D object detection. The proposed approach incorporates contrastive learning to enhance feature consistency between camera and LiDAR candidates, which is named as Contrastive Camera-LiDAR Object Candidates (C-CLOCs) fusion network, facilitating better fusion results. We delve into the label assignment aspect in late fusion methods and introduce a novel label assignment strategy to filter out irrelevant information. Additionally, a Multi-modality Ground-truth Sampling (MGS) method is introduced, which leverages the inclusion of point cloud information from LiDAR and corresponding images in training samples, resulting in improved performance. Experimental results demonstrate the effectiveness of the proposed method in achieving accurate 3D object detection in autonomous driving scenarios.\",\"PeriodicalId\":36532,\"journal\":{\"name\":\"IEEE Transactions on Intelligent Vehicles\",\"volume\":\"10 5\",\"pages\":\"3442-3457\"},\"PeriodicalIF\":14.3000,\"publicationDate\":\"2024-09-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Intelligent Vehicles\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10663866/\",\"RegionNum\":1,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Intelligent Vehicles","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10663866/","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
摘要
在自动驾驶领域,准确、高效的三维物体检测是确保安全、可靠运行的关键。本文主要研究了相机和激光雷达数据的后期融合,用于三维目标检测。该方法结合对比学习增强摄像机和LiDAR候选对象之间的特征一致性,称为对比摄像机-LiDAR候选对象融合网络(contrast camera -LiDAR Object candidate, C-CLOCs),融合效果更好。深入研究了后期融合方法中的标签分配问题,提出了一种新的标签分配策略来过滤掉不相关信息。此外,介绍了一种多模态地面真值采样(multi -modal Ground-truth Sampling, MGS)方法,该方法利用了在训练样本中包含LiDAR点云信息和相应图像,从而提高了性能。实验结果证明了该方法在自动驾驶场景下实现精确三维目标检测的有效性。
In the field of autonomous driving, accurate and efficient 3D object detection is crucial for ensuring safe and reliable operation. This paper focuses on the fusion of camera and LiDAR data in a late-fusion manner for 3D object detection. The proposed approach incorporates contrastive learning to enhance feature consistency between camera and LiDAR candidates, which is named as Contrastive Camera-LiDAR Object Candidates (C-CLOCs) fusion network, facilitating better fusion results. We delve into the label assignment aspect in late fusion methods and introduce a novel label assignment strategy to filter out irrelevant information. Additionally, a Multi-modality Ground-truth Sampling (MGS) method is introduced, which leverages the inclusion of point cloud information from LiDAR and corresponding images in training samples, resulting in improved performance. Experimental results demonstrate the effectiveness of the proposed method in achieving accurate 3D object detection in autonomous driving scenarios.
期刊介绍:
The IEEE Transactions on Intelligent Vehicles (T-IV) is a premier platform for publishing peer-reviewed articles that present innovative research concepts, application results, significant theoretical findings, and application case studies in the field of intelligent vehicles. With a particular emphasis on automated vehicles within roadway environments, T-IV aims to raise awareness of pressing research and application challenges.
Our focus is on providing critical information to the intelligent vehicle community, serving as a dissemination vehicle for IEEE ITS Society members and others interested in learning about the state-of-the-art developments and progress in research and applications related to intelligent vehicles. Join us in advancing knowledge and innovation in this dynamic field.