Hao Wang , Yongchao Shan , Liping Chen , Mengnan Liu , Lin Wang , Zhijun Meng
{"title":"基于无人机点云的农田三维语义映射多尺度特征学习","authors":"Hao Wang , Yongchao Shan , Liping Chen , Mengnan Liu , Lin Wang , Zhijun Meng","doi":"10.1016/j.jag.2025.104626","DOIUrl":null,"url":null,"abstract":"<div><div>Accurate spatial distribution information of field features is critical for enabling autonomous agricultural machinery navigation. However, current perception systems exhibit limited segmentation performance in complex farm environments due to illumination variations and mutual occlusion among various regions. This paper proposes a low-cost UAV photogrammetry framework for centimeter-level 3D semantic maps of agricultural fields to support autonomous agricultural machinery path planning. The methodology combines UAV-captured images with RTK positioning to reconstruct high-precision 3D point clouds, followed by a novel Local-Global Feature Aggregation Network (LoGA-Net) integrating multi-scale attention mechanisms and geometric constraints. The framework achieves 78.6% mIoU in classifying eight critical agricultural categories: paddy field, dry field, building, vegetation, farm track, paved ground, infrastructure and other static obstacles. Experimental validation demonstrates a 5.9% accuracy improvement over RandLA-Net on the Semantic3D benchmark. This advancement significantly enhances perception accuracy in complex agricultural environments, particularly for field boundary delineation and occluded feature recognition, which directly facilitates robust path planning for unmanned agricultural machinery. The framework provides a scalable technical and data-driven foundation for achieving fully autonomous farm operations, ensuring both operational efficiency and environmental sustainability.</div></div>","PeriodicalId":73423,"journal":{"name":"International journal of applied earth observation and geoinformation : ITC journal","volume":"141 ","pages":"Article 104626"},"PeriodicalIF":7.6000,"publicationDate":"2025-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-scale feature learning for 3D semantic mapping of agricultural fields using UAV point clouds\",\"authors\":\"Hao Wang , Yongchao Shan , Liping Chen , Mengnan Liu , Lin Wang , Zhijun Meng\",\"doi\":\"10.1016/j.jag.2025.104626\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Accurate spatial distribution information of field features is critical for enabling autonomous agricultural machinery navigation. However, current perception systems exhibit limited segmentation performance in complex farm environments due to illumination variations and mutual occlusion among various regions. This paper proposes a low-cost UAV photogrammetry framework for centimeter-level 3D semantic maps of agricultural fields to support autonomous agricultural machinery path planning. The methodology combines UAV-captured images with RTK positioning to reconstruct high-precision 3D point clouds, followed by a novel Local-Global Feature Aggregation Network (LoGA-Net) integrating multi-scale attention mechanisms and geometric constraints. The framework achieves 78.6% mIoU in classifying eight critical agricultural categories: paddy field, dry field, building, vegetation, farm track, paved ground, infrastructure and other static obstacles. Experimental validation demonstrates a 5.9% accuracy improvement over RandLA-Net on the Semantic3D benchmark. This advancement significantly enhances perception accuracy in complex agricultural environments, particularly for field boundary delineation and occluded feature recognition, which directly facilitates robust path planning for unmanned agricultural machinery. The framework provides a scalable technical and data-driven foundation for achieving fully autonomous farm operations, ensuring both operational efficiency and environmental sustainability.</div></div>\",\"PeriodicalId\":73423,\"journal\":{\"name\":\"International journal of applied earth observation and geoinformation : ITC journal\",\"volume\":\"141 \",\"pages\":\"Article 104626\"},\"PeriodicalIF\":7.6000,\"publicationDate\":\"2025-05-29\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International journal of applied earth observation and geoinformation : ITC journal\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1569843225002730\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"REMOTE SENSING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International journal of applied earth observation and geoinformation : ITC journal","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1569843225002730","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"REMOTE SENSING","Score":null,"Total":0}
Multi-scale feature learning for 3D semantic mapping of agricultural fields using UAV point clouds
Accurate spatial distribution information of field features is critical for enabling autonomous agricultural machinery navigation. However, current perception systems exhibit limited segmentation performance in complex farm environments due to illumination variations and mutual occlusion among various regions. This paper proposes a low-cost UAV photogrammetry framework for centimeter-level 3D semantic maps of agricultural fields to support autonomous agricultural machinery path planning. The methodology combines UAV-captured images with RTK positioning to reconstruct high-precision 3D point clouds, followed by a novel Local-Global Feature Aggregation Network (LoGA-Net) integrating multi-scale attention mechanisms and geometric constraints. The framework achieves 78.6% mIoU in classifying eight critical agricultural categories: paddy field, dry field, building, vegetation, farm track, paved ground, infrastructure and other static obstacles. Experimental validation demonstrates a 5.9% accuracy improvement over RandLA-Net on the Semantic3D benchmark. This advancement significantly enhances perception accuracy in complex agricultural environments, particularly for field boundary delineation and occluded feature recognition, which directly facilitates robust path planning for unmanned agricultural machinery. The framework provides a scalable technical and data-driven foundation for achieving fully autonomous farm operations, ensuring both operational efficiency and environmental sustainability.
期刊介绍:
The International Journal of Applied Earth Observation and Geoinformation publishes original papers that utilize earth observation data for natural resource and environmental inventory and management. These data primarily originate from remote sensing platforms, including satellites and aircraft, supplemented by surface and subsurface measurements. Addressing natural resources such as forests, agricultural land, soils, and water, as well as environmental concerns like biodiversity, land degradation, and hazards, the journal explores conceptual and data-driven approaches. It covers geoinformation themes like capturing, databasing, visualization, interpretation, data quality, and spatial uncertainty.