MMSeg:一种用于复杂野外环境中可导航区域的多模态多尺度点云分割模型

IF 5.2 2区 计算机科学 Q1 AUTOMATION & CONTROL SYSTEMS
Yifang Huang, Hongdou He, Peng Shi, Xiaobing Hao, Haitao He, Pei Miao
{"title":"MMSeg:一种用于复杂野外环境中可导航区域的多模态多尺度点云分割模型","authors":"Yifang Huang,&nbsp;Hongdou He,&nbsp;Peng Shi,&nbsp;Xiaobing Hao,&nbsp;Haitao He,&nbsp;Pei Miao","doi":"10.1016/j.robot.2025.105229","DOIUrl":null,"url":null,"abstract":"<div><div>This study focuses on advanced navigable area perception technology to address critical applications such as battlefield support and emergency rescue for autonomous ground intelligent agents. It emphasizes its application in complex field environments characterized by unstructured, diverse, and intricate features. Current methods predominantly target structured environments, neglecting the unique challenges presented by unstructured terrains essential for critical applications such as battlefield support and emergency rescue missions. To address this gap, we propose three key contributions through a Multimodal Multi-scale point cloud Segmentation (MMSeg) model. First, we introduce a multimodal ground feature fusion technique that integrates geometric information from LiDAR point clouds with visual texture features from images, enhancing the recognition capabilities of heterogeneous ground surfaces. Second, we propose a local–global terrain geometry information enhancement method that utilizes a dual-attention mechanism to effectively capture and analyze both local and global geometric features in complex terrain conditions. Third, we design a multi-scale classifier framework that effectively processes the multimodal fused information of ground materials and terrain structures, enabling precise segmentation of navigable areas. An experiment on a dedicated platform demonstrates that the MMSeg model achieves mIoU 6% higher than commonly used point cloud segmentation models. These findings suggest that the MMSeg model significantly enhances the perception capabilities of autonomous ground intelligent agents in challenging environments, providing a promising and novel solution to improve their operational effectiveness in complex field conditions.</div></div>","PeriodicalId":49592,"journal":{"name":"Robotics and Autonomous Systems","volume":"195 ","pages":"Article 105229"},"PeriodicalIF":5.2000,"publicationDate":"2025-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"MMSeg: A multimodal multi-scale point cloud segmentation model for navigable areas in complex field environments\",\"authors\":\"Yifang Huang,&nbsp;Hongdou He,&nbsp;Peng Shi,&nbsp;Xiaobing Hao,&nbsp;Haitao He,&nbsp;Pei Miao\",\"doi\":\"10.1016/j.robot.2025.105229\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>This study focuses on advanced navigable area perception technology to address critical applications such as battlefield support and emergency rescue for autonomous ground intelligent agents. It emphasizes its application in complex field environments characterized by unstructured, diverse, and intricate features. Current methods predominantly target structured environments, neglecting the unique challenges presented by unstructured terrains essential for critical applications such as battlefield support and emergency rescue missions. To address this gap, we propose three key contributions through a Multimodal Multi-scale point cloud Segmentation (MMSeg) model. First, we introduce a multimodal ground feature fusion technique that integrates geometric information from LiDAR point clouds with visual texture features from images, enhancing the recognition capabilities of heterogeneous ground surfaces. Second, we propose a local–global terrain geometry information enhancement method that utilizes a dual-attention mechanism to effectively capture and analyze both local and global geometric features in complex terrain conditions. Third, we design a multi-scale classifier framework that effectively processes the multimodal fused information of ground materials and terrain structures, enabling precise segmentation of navigable areas. An experiment on a dedicated platform demonstrates that the MMSeg model achieves mIoU 6% higher than commonly used point cloud segmentation models. These findings suggest that the MMSeg model significantly enhances the perception capabilities of autonomous ground intelligent agents in challenging environments, providing a promising and novel solution to improve their operational effectiveness in complex field conditions.</div></div>\",\"PeriodicalId\":49592,\"journal\":{\"name\":\"Robotics and Autonomous Systems\",\"volume\":\"195 \",\"pages\":\"Article 105229\"},\"PeriodicalIF\":5.2000,\"publicationDate\":\"2025-10-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Robotics and Autonomous Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0921889025003264\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Robotics and Autonomous Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0921889025003264","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

摘要

本研究的重点是先进的可导航区域感知技术,以解决自主地面智能体的战场支援和紧急救援等关键应用。强调其在非结构化、多样化、错综复杂的复杂野外环境中的应用。目前的方法主要针对结构化环境,忽视了非结构化地形所带来的独特挑战,而非结构化地形对于战场支援和紧急救援任务等关键应用至关重要。为了解决这一差距,我们通过多模态多尺度点云分割(MMSeg)模型提出了三个关键贡献。首先,我们引入了一种多模态地面特征融合技术,该技术将激光雷达点云的几何信息与图像的视觉纹理特征相结合,增强了对非均匀地面表面的识别能力。其次,我们提出了一种局部-全局地形几何信息增强方法,该方法利用双注意机制有效捕获和分析复杂地形条件下的局部和全局几何特征。第三,设计了一种多尺度分类器框架,有效地处理了地面材料和地形结构的多模态融合信息,实现了可通航区域的精确分割。在专用平台上的实验表明,MMSeg模型的mIoU比常用的点云分割模型高6%。这些发现表明,MMSeg模型显著增强了自主地面智能体在挑战性环境中的感知能力,为提高其在复杂野外条件下的作战效率提供了一种有前景的新颖解决方案。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
MMSeg: A multimodal multi-scale point cloud segmentation model for navigable areas in complex field environments
This study focuses on advanced navigable area perception technology to address critical applications such as battlefield support and emergency rescue for autonomous ground intelligent agents. It emphasizes its application in complex field environments characterized by unstructured, diverse, and intricate features. Current methods predominantly target structured environments, neglecting the unique challenges presented by unstructured terrains essential for critical applications such as battlefield support and emergency rescue missions. To address this gap, we propose three key contributions through a Multimodal Multi-scale point cloud Segmentation (MMSeg) model. First, we introduce a multimodal ground feature fusion technique that integrates geometric information from LiDAR point clouds with visual texture features from images, enhancing the recognition capabilities of heterogeneous ground surfaces. Second, we propose a local–global terrain geometry information enhancement method that utilizes a dual-attention mechanism to effectively capture and analyze both local and global geometric features in complex terrain conditions. Third, we design a multi-scale classifier framework that effectively processes the multimodal fused information of ground materials and terrain structures, enabling precise segmentation of navigable areas. An experiment on a dedicated platform demonstrates that the MMSeg model achieves mIoU 6% higher than commonly used point cloud segmentation models. These findings suggest that the MMSeg model significantly enhances the perception capabilities of autonomous ground intelligent agents in challenging environments, providing a promising and novel solution to improve their operational effectiveness in complex field conditions.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Robotics and Autonomous Systems
Robotics and Autonomous Systems 工程技术-机器人学
CiteScore
9.00
自引率
7.00%
发文量
164
审稿时长
4.5 months
期刊介绍: Robotics and Autonomous Systems will carry articles describing fundamental developments in the field of robotics, with special emphasis on autonomous systems. An important goal of this journal is to extend the state of the art in both symbolic and sensory based robot control and learning in the context of autonomous systems. Robotics and Autonomous Systems will carry articles on the theoretical, computational and experimental aspects of autonomous systems, or modules of such systems.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信