Yudi Ruan , Hao Ma , Di Ma , Weikai Li , Xiao Wang
{"title":"利用双交叉注意增强弱光图像","authors":"Yudi Ruan , Hao Ma , Di Ma , Weikai Li , Xiao Wang","doi":"10.1016/j.engappai.2025.111501","DOIUrl":null,"url":null,"abstract":"<div><div>Low-light image enhancement (LLIE) aims to improve the perceptibility and interpretability of images captured in poorly illuminated environments. Existing LLIE methods often fail to capture the local self-similarity and long-range dependencies at the same time, causing the loss of complementary information between multiple modules or network layers, ultimately resulting in the loss of image details. To alleviate this issue, we design a hierarchical mutual Enhancement via a dual cross-attention transformer (ECAFormer), which introduces an architecture that enables concurrent propagation and interaction of multiple disentangling features. To capture the local self-similarity, we design a Dual Multi-head Self-Attention (DMSA), which leverages the disentangled visual and semantic features across different scales, allowing them to guide and complement each other. Further, a cross-scale DMSA block is incorporated to capture residual connections, thereby integrating cross-layer information and capturing the long-range dependencies. Experimental results show that the ECAFormer reaches competitive performance across multiple benchmarks, yielding nearly a 3.7% improvement in Peak Signal-to-Noise Ratio (PSNR) over the suboptimal method, demonstrating the effectiveness of information interaction in LLIE. For facilitating the efforts to replicate our results, our implementation is available on GitHub<span><span><sup>1</sup></span></span></div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"159 ","pages":"Article 111501"},"PeriodicalIF":8.0000,"publicationDate":"2025-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Low-light image enhancement using dual cross attention\",\"authors\":\"Yudi Ruan , Hao Ma , Di Ma , Weikai Li , Xiao Wang\",\"doi\":\"10.1016/j.engappai.2025.111501\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Low-light image enhancement (LLIE) aims to improve the perceptibility and interpretability of images captured in poorly illuminated environments. Existing LLIE methods often fail to capture the local self-similarity and long-range dependencies at the same time, causing the loss of complementary information between multiple modules or network layers, ultimately resulting in the loss of image details. To alleviate this issue, we design a hierarchical mutual Enhancement via a dual cross-attention transformer (ECAFormer), which introduces an architecture that enables concurrent propagation and interaction of multiple disentangling features. To capture the local self-similarity, we design a Dual Multi-head Self-Attention (DMSA), which leverages the disentangled visual and semantic features across different scales, allowing them to guide and complement each other. Further, a cross-scale DMSA block is incorporated to capture residual connections, thereby integrating cross-layer information and capturing the long-range dependencies. Experimental results show that the ECAFormer reaches competitive performance across multiple benchmarks, yielding nearly a 3.7% improvement in Peak Signal-to-Noise Ratio (PSNR) over the suboptimal method, demonstrating the effectiveness of information interaction in LLIE. For facilitating the efforts to replicate our results, our implementation is available on GitHub<span><span><sup>1</sup></span></span></div></div>\",\"PeriodicalId\":50523,\"journal\":{\"name\":\"Engineering Applications of Artificial Intelligence\",\"volume\":\"159 \",\"pages\":\"Article 111501\"},\"PeriodicalIF\":8.0000,\"publicationDate\":\"2025-07-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Engineering Applications of Artificial Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0952197625015039\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197625015039","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
Low-light image enhancement using dual cross attention
Low-light image enhancement (LLIE) aims to improve the perceptibility and interpretability of images captured in poorly illuminated environments. Existing LLIE methods often fail to capture the local self-similarity and long-range dependencies at the same time, causing the loss of complementary information between multiple modules or network layers, ultimately resulting in the loss of image details. To alleviate this issue, we design a hierarchical mutual Enhancement via a dual cross-attention transformer (ECAFormer), which introduces an architecture that enables concurrent propagation and interaction of multiple disentangling features. To capture the local self-similarity, we design a Dual Multi-head Self-Attention (DMSA), which leverages the disentangled visual and semantic features across different scales, allowing them to guide and complement each other. Further, a cross-scale DMSA block is incorporated to capture residual connections, thereby integrating cross-layer information and capturing the long-range dependencies. Experimental results show that the ECAFormer reaches competitive performance across multiple benchmarks, yielding nearly a 3.7% improvement in Peak Signal-to-Noise Ratio (PSNR) over the suboptimal method, demonstrating the effectiveness of information interaction in LLIE. For facilitating the efforts to replicate our results, our implementation is available on GitHub1
期刊介绍:
Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.