RGBD Semantic Segmentation Based on Global Convolutional Network

Xiaoning Gao, Jun Yu, Jian-xun Li
{"title":"RGBD Semantic Segmentation Based on Global Convolutional Network","authors":"Xiaoning Gao, Jun Yu, Jian-xun Li","doi":"10.1145/3351180.3351182","DOIUrl":null,"url":null,"abstract":"Convolutional neural networks have gradually dominated the field of image semantic segmentation, and have achieved good results in 2D image semantic segmentation tasks. However, the 2D semantic segmentation algorithm based on CNN is still unsatisfactory in many complex scenarios, such as indoor scenes. Fortunately, advances in depth sensor technology have made it easy to obtain depth information, which carries rich geometric structure information. In order to effectively embed the depth map into the convolutional neural network, this paper introduces the dual encoder fusion network framework to fully exploit the geometric features. For the problem of weakening the local pixel classification ability of the dual encoder fusion network, this paper introduces global convolutional network (GCN), which is based on the large kernel idea, to improve the performance of dual encoder fusion network. Extensive experiments on the NYU v2 dataset demonstrate that the two-encoder fusion network based on global convolution network has much better precision than the original fusion network, and the classification ability for local pixels is stronger.","PeriodicalId":375806,"journal":{"name":"Proceedings of the 2019 4th International Conference on Robotics, Control and Automation","volume":"3 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2019 4th International Conference on Robotics, Control and Automation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3351180.3351182","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

Convolutional neural networks have gradually dominated the field of image semantic segmentation, and have achieved good results in 2D image semantic segmentation tasks. However, the 2D semantic segmentation algorithm based on CNN is still unsatisfactory in many complex scenarios, such as indoor scenes. Fortunately, advances in depth sensor technology have made it easy to obtain depth information, which carries rich geometric structure information. In order to effectively embed the depth map into the convolutional neural network, this paper introduces the dual encoder fusion network framework to fully exploit the geometric features. For the problem of weakening the local pixel classification ability of the dual encoder fusion network, this paper introduces global convolutional network (GCN), which is based on the large kernel idea, to improve the performance of dual encoder fusion network. Extensive experiments on the NYU v2 dataset demonstrate that the two-encoder fusion network based on global convolution network has much better precision than the original fusion network, and the classification ability for local pixels is stronger.
基于全局卷积网络的RGBD语义分割
卷积神经网络逐渐主导了图像语义分割领域,并在二维图像语义分割任务中取得了良好的效果。然而,基于CNN的二维语义分割算法在室内场景等许多复杂场景下仍然不尽人意。幸运的是,深度传感器技术的进步使得深度信息的获取变得容易,而深度信息承载着丰富的几何结构信息。为了有效地将深度图嵌入到卷积神经网络中,本文引入了双编码器融合网络框架,充分利用深度图的几何特征。针对双编码器融合网络局部像素分类能力减弱的问题,本文引入基于大核思想的全局卷积网络(global convolutional network, GCN)来提高双编码器融合网络的性能。在NYU v2数据集上的大量实验表明,基于全局卷积网络的双编码器融合网络比原始融合网络具有更好的精度,并且对局部像素的分类能力更强。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信