Semantic Segmentation Based on Deeplabv3+ and Attention Mechanism

Rongrong Liu, Dongzhi He
{"title":"Semantic Segmentation Based on Deeplabv3+ and Attention Mechanism","authors":"Rongrong Liu, Dongzhi He","doi":"10.1109/IMCEC51613.2021.9482207","DOIUrl":null,"url":null,"abstract":"In this paper, we propose vertical attention and spatial attention network (VSANet), which is a semantic segmentation method based on Deeplabv3+ and attention module, for improving semantic segmentation effect for autonomous driving road scene images. The improvement of this paper is primarily on the following two aspects. One is that this paper introduces the spatial attention module (SAM) after the atrous convolution, which effectively obtains more spatial context information. Second, by studying the road scene image, it is found that there are considerable differences in the pixel-level distribution of the horizontal segmentation area in the image. For this reason, this paper introduces the vertical attention module (VAM), which can better segment the road scene image. A large number of experimental results indicate that the segmentation accuracy of the proposed model is improved by 1.94% compared with the Deeplabv3+ network model on the test dataset of Cityscapes dataset.","PeriodicalId":240400,"journal":{"name":"2021 IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC)","volume":"142 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IMCEC51613.2021.9482207","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

In this paper, we propose vertical attention and spatial attention network (VSANet), which is a semantic segmentation method based on Deeplabv3+ and attention module, for improving semantic segmentation effect for autonomous driving road scene images. The improvement of this paper is primarily on the following two aspects. One is that this paper introduces the spatial attention module (SAM) after the atrous convolution, which effectively obtains more spatial context information. Second, by studying the road scene image, it is found that there are considerable differences in the pixel-level distribution of the horizontal segmentation area in the image. For this reason, this paper introduces the vertical attention module (VAM), which can better segment the road scene image. A large number of experimental results indicate that the segmentation accuracy of the proposed model is improved by 1.94% compared with the Deeplabv3+ network model on the test dataset of Cityscapes dataset.
基于Deeplabv3+和注意机制的语义分割
为了提高自动驾驶道路场景图像的语义分割效果,本文提出了一种基于Deeplabv3+和注意力模块的语义分割方法——垂直注意力和空间注意力网络(vertical attention and spatial attention network, VSANet)。本文的改进主要体现在以下两个方面。一是在亚历斯卷积之后引入空间注意模块(SAM),有效地获取了更多的空间上下文信息。其次,通过对道路场景图像的研究,发现图像中水平分割区域的像素级分布存在较大差异。为此,本文引入了垂直关注模块(vertical attention module, VAM),该模块可以更好地分割道路场景图像。大量实验结果表明,在cityscape数据集的测试数据集上,与Deeplabv3+网络模型相比,该模型的分割精度提高了1.94%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信