{"title":"基于交叉注意机制的多曝光图像融合","authors":"Byungnam Kim, Hyungjoo Jung, K. Sohn","doi":"10.1109/ICCE53296.2022.9730511","DOIUrl":null,"url":null,"abstract":"Multi-exposure fusion (MEF) is a popular method for obtaining high dynamic range (HDR) image from multiple low dynamic range (LDR) images. Even though recent works have employed the convolutional neural networks (CNNs) for solving the MEF problem, there still remain various challenges, such as color distortion and detail loss, due to a limited receptive field. In this paper, we present a cross-attention module for multi-exposure image fusion. Different from existing CNN-based methods that capture the contexts of the local region in the target image, our method adaptively aggregates local features with global dependencies at all positions. Furthermore, we propose a detail compensation module as the feature fusion for restoring the loss (color and detail) in the saturation region. Our proposed network performs a feature extraction with an encoder, a fusion of a cross-attention module and a detail compensation module, and the fused image is reconstructed by a decoder. Experimental results show that compared with the state-of-the-art methods, the proposed method can obtain better performance in both the subjective and objective evaluation, particularly in terms of color expression and detail-preserving.","PeriodicalId":350644,"journal":{"name":"2022 IEEE International Conference on Consumer Electronics (ICCE)","volume":"78 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Multi-Exposure Image Fusion Using Cross-Attention Mechanism\",\"authors\":\"Byungnam Kim, Hyungjoo Jung, K. Sohn\",\"doi\":\"10.1109/ICCE53296.2022.9730511\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multi-exposure fusion (MEF) is a popular method for obtaining high dynamic range (HDR) image from multiple low dynamic range (LDR) images. Even though recent works have employed the convolutional neural networks (CNNs) for solving the MEF problem, there still remain various challenges, such as color distortion and detail loss, due to a limited receptive field. In this paper, we present a cross-attention module for multi-exposure image fusion. Different from existing CNN-based methods that capture the contexts of the local region in the target image, our method adaptively aggregates local features with global dependencies at all positions. Furthermore, we propose a detail compensation module as the feature fusion for restoring the loss (color and detail) in the saturation region. Our proposed network performs a feature extraction with an encoder, a fusion of a cross-attention module and a detail compensation module, and the fused image is reconstructed by a decoder. Experimental results show that compared with the state-of-the-art methods, the proposed method can obtain better performance in both the subjective and objective evaluation, particularly in terms of color expression and detail-preserving.\",\"PeriodicalId\":350644,\"journal\":{\"name\":\"2022 IEEE International Conference on Consumer Electronics (ICCE)\",\"volume\":\"78 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-01-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE International Conference on Consumer Electronics (ICCE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCE53296.2022.9730511\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE International Conference on Consumer Electronics (ICCE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCE53296.2022.9730511","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multi-Exposure Image Fusion Using Cross-Attention Mechanism
Multi-exposure fusion (MEF) is a popular method for obtaining high dynamic range (HDR) image from multiple low dynamic range (LDR) images. Even though recent works have employed the convolutional neural networks (CNNs) for solving the MEF problem, there still remain various challenges, such as color distortion and detail loss, due to a limited receptive field. In this paper, we present a cross-attention module for multi-exposure image fusion. Different from existing CNN-based methods that capture the contexts of the local region in the target image, our method adaptively aggregates local features with global dependencies at all positions. Furthermore, we propose a detail compensation module as the feature fusion for restoring the loss (color and detail) in the saturation region. Our proposed network performs a feature extraction with an encoder, a fusion of a cross-attention module and a detail compensation module, and the fused image is reconstructed by a decoder. Experimental results show that compared with the state-of-the-art methods, the proposed method can obtain better performance in both the subjective and objective evaluation, particularly in terms of color expression and detail-preserving.