一种用于深度高动态范围成像的新型注意引导网络

Qinghan Jiang, Ying Huang, Su Liu, Zequan Wang, Tangsheng Li
{"title":"一种用于深度高动态范围成像的新型注意引导网络","authors":"Qinghan Jiang, Ying Huang, Su Liu, Zequan Wang, Tangsheng Li","doi":"10.1109/CTISC52352.2021.00069","DOIUrl":null,"url":null,"abstract":"In natural scenes with multi-exposure image fusion (MEF), high dynamic range (HDR) imaging is often affected by moving objects or misalignments in the scene, resulting in ghosting artifacts in the final imaging results, with the help of optical flow method and deep network architecture. To avoid ghosting artifacts better, we propose a novel attention- guided neural network (ADeepHDR) to produce high-quality ghost-free HDR images. Unlike the previous methods, we use the attention module to guide the process of image merging. The attention module can detect the large motions and the notable parts of the different input features and enhance details in the results. Based on the attention module, we also try different subnetwork variants to make full use of the hierarchical features to get more ideal results. Besides, fractional-oder differential convolution is used in the subnetwork variant to extract more detailed features. The proposed ADeepHDR is an improvement method without optical flows, which can better avoid the ghosting artifacts caused by error optical flow estimation and large motions. We have conducted extensive quantitative and qualitative assessments, and show that the proposed method is superior to the most state-of-the- art approaches.","PeriodicalId":268378,"journal":{"name":"2021 3rd International Conference on Advances in Computer Technology, Information Science and Communication (CTISC)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Novel Attention-guided Network for Deep High Dynamic Range Imaging\",\"authors\":\"Qinghan Jiang, Ying Huang, Su Liu, Zequan Wang, Tangsheng Li\",\"doi\":\"10.1109/CTISC52352.2021.00069\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In natural scenes with multi-exposure image fusion (MEF), high dynamic range (HDR) imaging is often affected by moving objects or misalignments in the scene, resulting in ghosting artifacts in the final imaging results, with the help of optical flow method and deep network architecture. To avoid ghosting artifacts better, we propose a novel attention- guided neural network (ADeepHDR) to produce high-quality ghost-free HDR images. Unlike the previous methods, we use the attention module to guide the process of image merging. The attention module can detect the large motions and the notable parts of the different input features and enhance details in the results. Based on the attention module, we also try different subnetwork variants to make full use of the hierarchical features to get more ideal results. Besides, fractional-oder differential convolution is used in the subnetwork variant to extract more detailed features. The proposed ADeepHDR is an improvement method without optical flows, which can better avoid the ghosting artifacts caused by error optical flow estimation and large motions. We have conducted extensive quantitative and qualitative assessments, and show that the proposed method is superior to the most state-of-the- art approaches.\",\"PeriodicalId\":268378,\"journal\":{\"name\":\"2021 3rd International Conference on Advances in Computer Technology, Information Science and Communication (CTISC)\",\"volume\":\"32 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-04-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 3rd International Conference on Advances in Computer Technology, Information Science and Communication (CTISC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CTISC52352.2021.00069\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 3rd International Conference on Advances in Computer Technology, Information Science and Communication (CTISC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CTISC52352.2021.00069","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

在多曝光图像融合(MEF)的自然场景中,高动态范围(HDR)成像经常受到场景中运动物体或错位的影响,导致最终成像结果出现重影伪影,借助光流方法和深度网络架构。为了更好地避免鬼影伪影,我们提出了一种新的注意力引导神经网络(ADeepHDR)来产生高质量的无鬼HDR图像。与之前的方法不同,我们使用注意力模块来指导图像合并过程。注意模块可以检测不同输入特征中较大的运动和值得注意的部分,并增强结果中的细节。在注意模块的基础上,我们还尝试了不同的子网变体,以充分利用分层特征,得到更理想的结果。此外,在子网变体中使用分数阶微分卷积来提取更详细的特征。本文提出的ADeepHDR是一种不含光流的改进方法,可以更好地避免光流估计误差和大运动引起的重影伪影。我们进行了广泛的定量和定性评估,并表明所提出的方法优于最先进的方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A Novel Attention-guided Network for Deep High Dynamic Range Imaging
In natural scenes with multi-exposure image fusion (MEF), high dynamic range (HDR) imaging is often affected by moving objects or misalignments in the scene, resulting in ghosting artifacts in the final imaging results, with the help of optical flow method and deep network architecture. To avoid ghosting artifacts better, we propose a novel attention- guided neural network (ADeepHDR) to produce high-quality ghost-free HDR images. Unlike the previous methods, we use the attention module to guide the process of image merging. The attention module can detect the large motions and the notable parts of the different input features and enhance details in the results. Based on the attention module, we also try different subnetwork variants to make full use of the hierarchical features to get more ideal results. Besides, fractional-oder differential convolution is used in the subnetwork variant to extract more detailed features. The proposed ADeepHDR is an improvement method without optical flows, which can better avoid the ghosting artifacts caused by error optical flow estimation and large motions. We have conducted extensive quantitative and qualitative assessments, and show that the proposed method is superior to the most state-of-the- art approaches.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信