{"title":"Joint Semantic-geometric Mapping of Unstructured Environment for Autonomous Mobile Robotic Sprayers","authors":"Xubin Lin, Zerong Su, Zhihan Zhu, Pengfei Yuan, Haifei Zhu, Xuefeng Zhou","doi":"10.1002/rob.22553","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>Mobile robotic sprayers are expected to be employed in outdoor insecticide applications for mosquito control, epidemic prevention, and disinfection. To achieve this, a comprehensive 3D environmental model integrating both semantic and geometric information is indispensable for supporting mobile robotic sprayers in autonomous navigation, task planning, and adaptive spraying control. However, outdoor environments for insecticide spraying, such as public parks and gardens, are typically unstructured, dynamic and prone to sensor degradation, posing significant challenges to both LiDAR-only and camera-only perception and mapping approaches. In this paper, a visual-LiDAR fusion based joint semantic-geometric mapping framework is proposed, featuring a novel 2D-3D semantic perception module that is robust against complex segmentation conditions and sensor extrinsic drift. To this end, a Multi-scale Vague Boundary Augmented Dual Attention Network (MDANet), incorporating multi-scale 3D attention modules and vague boundary augmented attention modules, is proposed to tackle the image segmentation task involving dense vegetation with overlapping foliage and ambiguous boundaries. Additionally, a seed growth-based visual-LiDAR semantic data association method is proposed to resolve the issue of inaccurate pixel-to-point association in the presence of extrinsic drift, yielding more precise 3D semantic perception results. Furthermore, a semantic-aware SLAM system accounting for dynamic interference and pose estimation drift is presented. Extensive experimental evaluations on public datasets and self-recorded data are conducted. The segmentation results show that MDANet achieves a mean pixel accuracy (mPA) of 90.17%, outperforming competing methods in the vegetation-involved segmentation task. The proposed visual-LiDAR semantic data association method can tolerate a translational disturbance of up to 40 mm and a rotational disturbance of 0.18 rad without compromising 3D segmentation accuracy. Moreover, the evaluation of trajectory error, alongside ablation studies, validates the effectiveness and feasibility of the proposed mapping framework.</p>\n </div>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"42 6","pages":"2952-2967"},"PeriodicalIF":5.2000,"publicationDate":"2025-04-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Field Robotics","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/rob.22553","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0
Abstract
Mobile robotic sprayers are expected to be employed in outdoor insecticide applications for mosquito control, epidemic prevention, and disinfection. To achieve this, a comprehensive 3D environmental model integrating both semantic and geometric information is indispensable for supporting mobile robotic sprayers in autonomous navigation, task planning, and adaptive spraying control. However, outdoor environments for insecticide spraying, such as public parks and gardens, are typically unstructured, dynamic and prone to sensor degradation, posing significant challenges to both LiDAR-only and camera-only perception and mapping approaches. In this paper, a visual-LiDAR fusion based joint semantic-geometric mapping framework is proposed, featuring a novel 2D-3D semantic perception module that is robust against complex segmentation conditions and sensor extrinsic drift. To this end, a Multi-scale Vague Boundary Augmented Dual Attention Network (MDANet), incorporating multi-scale 3D attention modules and vague boundary augmented attention modules, is proposed to tackle the image segmentation task involving dense vegetation with overlapping foliage and ambiguous boundaries. Additionally, a seed growth-based visual-LiDAR semantic data association method is proposed to resolve the issue of inaccurate pixel-to-point association in the presence of extrinsic drift, yielding more precise 3D semantic perception results. Furthermore, a semantic-aware SLAM system accounting for dynamic interference and pose estimation drift is presented. Extensive experimental evaluations on public datasets and self-recorded data are conducted. The segmentation results show that MDANet achieves a mean pixel accuracy (mPA) of 90.17%, outperforming competing methods in the vegetation-involved segmentation task. The proposed visual-LiDAR semantic data association method can tolerate a translational disturbance of up to 40 mm and a rotational disturbance of 0.18 rad without compromising 3D segmentation accuracy. Moreover, the evaluation of trajectory error, alongside ablation studies, validates the effectiveness and feasibility of the proposed mapping framework.
期刊介绍:
The Journal of Field Robotics seeks to promote scholarly publications dealing with the fundamentals of robotics in unstructured and dynamic environments.
The Journal focuses on experimental robotics and encourages publication of work that has both theoretical and practical significance.