Hao Lu, Bowen Li, Gang Yang, Guangpeng Fan, Han Wang, Yong Pang, Zuyuan Wang, Yining Lian, Haifeng Xu, Huan Huang
{"title":"面向森林类型和传感器平台的森林场景语义分割的点云理解框架","authors":"Hao Lu, Bowen Li, Gang Yang, Guangpeng Fan, Han Wang, Yong Pang, Zuyuan Wang, Yining Lian, Haifeng Xu, Huan Huang","doi":"10.1016/j.rse.2024.114591","DOIUrl":null,"url":null,"abstract":"Foliage, wood (i.e., trunk, branch), ground, and lower objects (i.e., grass, shrubs), are key semantic components of forests that play different roles in the forest ecosystem. For understanding forest ecosystem structure and function, Light Detection And Ranging (LiDAR) point cloud is a valuable form of remote sensing observation. The understanding intensively relies on precisely performing semantic segmentation task to segment semantic components from forest point cloud data. However, the semantic segmentation of massive point cloud data from forest scenes remains a significant challenge. The forest environment is highly heterogeneous and complex due to tree species and terrain conditions. Different climate zones lead to varying canopy characteristics and the diversity of LiDAR platforms delivers inconsistent point cloud properties. Heuristic approaches and conventional machine learning approaches inevitably suffer from poor generalization. Additionally, most deep learning methods lack a dedicated network design to address the characteristics of forests. This paper introduces Sen-net, a point cloud understanding network specifically constructed for semantic component segmentation of forest scene point cloud. Sen-net implements three modules tailored for forest characteristics. First, a spatial context enhancement module (SCEM) is designed for providing both global and dataset-level perspectives to mine geometric information and robust features hidden in heterogeneous forest. Second, a semantic-driven detail enrichment module (SDEM) is incorporated to preserve rich geometric details and semantic information thereby enhancing the learning of complex structures in the forests. Finally, an adaptive guidance flow (AGF) is added to seamlessly fuse the semantic and detailed features. Comprehensive experiments were conducted on both self-built Lin3D dataset and public datasets. Sen-net achieved an OA of 97.6 % and 85.1 % MIoU on the Lin3D dataset, and an OA of 94.5 % and 78.2 % MIoU on the public dataset FOR-instance. Results show that Sen-net outperformed the representative forest scene point cloud semantic segmentation approaches and state-of-the-art deep learning networks, and it has the potential to generalize to point cloud data collected by LiDAR from other platforms. It is concluded that Sen-net is a powerful and robust framework with substantial potential for being widely and deeply explored in forest ecosystem studies.","PeriodicalId":417,"journal":{"name":"Remote Sensing of Environment","volume":"22 1","pages":""},"PeriodicalIF":11.1000,"publicationDate":"2025-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Towards a point cloud understanding framework for forest scene semantic segmentation across forest types and sensor platforms\",\"authors\":\"Hao Lu, Bowen Li, Gang Yang, Guangpeng Fan, Han Wang, Yong Pang, Zuyuan Wang, Yining Lian, Haifeng Xu, Huan Huang\",\"doi\":\"10.1016/j.rse.2024.114591\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Foliage, wood (i.e., trunk, branch), ground, and lower objects (i.e., grass, shrubs), are key semantic components of forests that play different roles in the forest ecosystem. For understanding forest ecosystem structure and function, Light Detection And Ranging (LiDAR) point cloud is a valuable form of remote sensing observation. The understanding intensively relies on precisely performing semantic segmentation task to segment semantic components from forest point cloud data. However, the semantic segmentation of massive point cloud data from forest scenes remains a significant challenge. The forest environment is highly heterogeneous and complex due to tree species and terrain conditions. Different climate zones lead to varying canopy characteristics and the diversity of LiDAR platforms delivers inconsistent point cloud properties. Heuristic approaches and conventional machine learning approaches inevitably suffer from poor generalization. Additionally, most deep learning methods lack a dedicated network design to address the characteristics of forests. This paper introduces Sen-net, a point cloud understanding network specifically constructed for semantic component segmentation of forest scene point cloud. Sen-net implements three modules tailored for forest characteristics. First, a spatial context enhancement module (SCEM) is designed for providing both global and dataset-level perspectives to mine geometric information and robust features hidden in heterogeneous forest. Second, a semantic-driven detail enrichment module (SDEM) is incorporated to preserve rich geometric details and semantic information thereby enhancing the learning of complex structures in the forests. Finally, an adaptive guidance flow (AGF) is added to seamlessly fuse the semantic and detailed features. Comprehensive experiments were conducted on both self-built Lin3D dataset and public datasets. Sen-net achieved an OA of 97.6 % and 85.1 % MIoU on the Lin3D dataset, and an OA of 94.5 % and 78.2 % MIoU on the public dataset FOR-instance. Results show that Sen-net outperformed the representative forest scene point cloud semantic segmentation approaches and state-of-the-art deep learning networks, and it has the potential to generalize to point cloud data collected by LiDAR from other platforms. It is concluded that Sen-net is a powerful and robust framework with substantial potential for being widely and deeply explored in forest ecosystem studies.\",\"PeriodicalId\":417,\"journal\":{\"name\":\"Remote Sensing of Environment\",\"volume\":\"22 1\",\"pages\":\"\"},\"PeriodicalIF\":11.1000,\"publicationDate\":\"2025-01-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Remote Sensing of Environment\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://doi.org/10.1016/j.rse.2024.114591\",\"RegionNum\":1,\"RegionCategory\":\"地球科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENVIRONMENTAL SCIENCES\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Remote Sensing of Environment","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1016/j.rse.2024.114591","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENVIRONMENTAL SCIENCES","Score":null,"Total":0}
Towards a point cloud understanding framework for forest scene semantic segmentation across forest types and sensor platforms
Foliage, wood (i.e., trunk, branch), ground, and lower objects (i.e., grass, shrubs), are key semantic components of forests that play different roles in the forest ecosystem. For understanding forest ecosystem structure and function, Light Detection And Ranging (LiDAR) point cloud is a valuable form of remote sensing observation. The understanding intensively relies on precisely performing semantic segmentation task to segment semantic components from forest point cloud data. However, the semantic segmentation of massive point cloud data from forest scenes remains a significant challenge. The forest environment is highly heterogeneous and complex due to tree species and terrain conditions. Different climate zones lead to varying canopy characteristics and the diversity of LiDAR platforms delivers inconsistent point cloud properties. Heuristic approaches and conventional machine learning approaches inevitably suffer from poor generalization. Additionally, most deep learning methods lack a dedicated network design to address the characteristics of forests. This paper introduces Sen-net, a point cloud understanding network specifically constructed for semantic component segmentation of forest scene point cloud. Sen-net implements three modules tailored for forest characteristics. First, a spatial context enhancement module (SCEM) is designed for providing both global and dataset-level perspectives to mine geometric information and robust features hidden in heterogeneous forest. Second, a semantic-driven detail enrichment module (SDEM) is incorporated to preserve rich geometric details and semantic information thereby enhancing the learning of complex structures in the forests. Finally, an adaptive guidance flow (AGF) is added to seamlessly fuse the semantic and detailed features. Comprehensive experiments were conducted on both self-built Lin3D dataset and public datasets. Sen-net achieved an OA of 97.6 % and 85.1 % MIoU on the Lin3D dataset, and an OA of 94.5 % and 78.2 % MIoU on the public dataset FOR-instance. Results show that Sen-net outperformed the representative forest scene point cloud semantic segmentation approaches and state-of-the-art deep learning networks, and it has the potential to generalize to point cloud data collected by LiDAR from other platforms. It is concluded that Sen-net is a powerful and robust framework with substantial potential for being widely and deeply explored in forest ecosystem studies.
期刊介绍:
Remote Sensing of Environment (RSE) serves the Earth observation community by disseminating results on the theory, science, applications, and technology that contribute to advancing the field of remote sensing. With a thoroughly interdisciplinary approach, RSE encompasses terrestrial, oceanic, and atmospheric sensing.
The journal emphasizes biophysical and quantitative approaches to remote sensing at local to global scales, covering a diverse range of applications and techniques.
RSE serves as a vital platform for the exchange of knowledge and advancements in the dynamic field of remote sensing.