Self Attention in U-Net for Semantic Segmentation of Low Resolution SAR Images

Hrishikesh Singh Yadav, Priyanshu Panchal, Divyanshu Manawat, G. S, S. S
{"title":"Self Attention in U-Net for Semantic Segmentation of Low Resolution SAR Images","authors":"Hrishikesh Singh Yadav, Priyanshu Panchal, Divyanshu Manawat, G. S, S. S","doi":"10.1145/3596286.3596291","DOIUrl":null,"url":null,"abstract":"The SAR image semantic segmentation using computer vision techniques has gained much popularity in the research community due to their wide applications. Despite the advancements in Deep Learning for image analysis, these models still struggle to segment SAR images due to the existence of speckle noise and a poor feature extractor. Moreover, deep learning models are challenging to train on small datasets and the performance of the model is significantly impacted by the quality of the data. This calls for the development of an effective network that can draw out critical information from the low resolution SAR images. In this regard, the present work proposes a unique Self attention module in U-Net for the semantic segmentation of low resolution SAR images.. The Self Attention Model makes use of Laplacian kernel to highlight the sharp discontinuities in the features that define the boundaries of the objects. The proposed model, employs dilated convolution layers at the initial layers, enabling the model to more effectively capture larger contextual information. With an accuracy of 0.84 and an F1-score of 0.83, the proposed model outperforms the state-of-the-art techniques in semantic segmentation of low resolution SAR images. The results clearly demonstrate the importance of the self attention module and the consideration of dilated convolution layers in the initial layers in semantic segmentation of low resolution SAR images.","PeriodicalId":208318,"journal":{"name":"Proceedings of the 2023 Asia Conference on Computer Vision, Image Processing and Pattern Recognition","volume":"50 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2023 Asia Conference on Computer Vision, Image Processing and Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3596286.3596291","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The SAR image semantic segmentation using computer vision techniques has gained much popularity in the research community due to their wide applications. Despite the advancements in Deep Learning for image analysis, these models still struggle to segment SAR images due to the existence of speckle noise and a poor feature extractor. Moreover, deep learning models are challenging to train on small datasets and the performance of the model is significantly impacted by the quality of the data. This calls for the development of an effective network that can draw out critical information from the low resolution SAR images. In this regard, the present work proposes a unique Self attention module in U-Net for the semantic segmentation of low resolution SAR images.. The Self Attention Model makes use of Laplacian kernel to highlight the sharp discontinuities in the features that define the boundaries of the objects. The proposed model, employs dilated convolution layers at the initial layers, enabling the model to more effectively capture larger contextual information. With an accuracy of 0.84 and an F1-score of 0.83, the proposed model outperforms the state-of-the-art techniques in semantic segmentation of low resolution SAR images. The results clearly demonstrate the importance of the self attention module and the consideration of dilated convolution layers in the initial layers in semantic segmentation of low resolution SAR images.
低分辨率SAR图像语义分割的U-Net自关注
基于计算机视觉的SAR图像语义分割技术因其广泛的应用而受到了研究界的广泛关注。尽管深度学习在图像分析方面取得了进步,但由于存在散斑噪声和较差的特征提取器,这些模型仍然难以分割SAR图像。此外,在小数据集上训练深度学习模型具有挑战性,并且模型的性能受到数据质量的显着影响。这就要求开发一个有效的网络,从低分辨率SAR图像中提取关键信息。在这方面,本研究在U-Net中提出了一个独特的自关注模块,用于低分辨率SAR图像的语义分割。自注意模型利用拉普拉斯核来突出定义物体边界的特征中的明显不连续。该模型在初始层采用扩展卷积层,使模型能够更有效地捕获更大的上下文信息。该模型的精度为0.84,f1分数为0.83,在低分辨率SAR图像的语义分割中优于最先进的技术。结果表明,在低分辨率SAR图像的语义分割中,自关注模块的重要性和初始层中扩展卷积层的考虑。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信