Guanhao Chen, Dan Xu, Kangjian He, Hongzhen Shi, Hao Zhang
{"title":"tsa - net:基于转置稀疏自关注的图像超分辨率网络","authors":"Guanhao Chen, Dan Xu, Kangjian He, Hongzhen Shi, Hao Zhang","doi":"10.1016/j.engappai.2025.110823","DOIUrl":null,"url":null,"abstract":"<div><div>Image super-resolution (SR) aims to reconstruct high-resolution images from low-resolution ones. Transformer-based methods have recently demonstrated remarkable results, but the conventional dense self-attention mechanism fails to capture local relationships between patches, leading to suboptimal performance. Moreover, the recovery of high-frequency information, crucial for edge reconstruction, is often insufficient. To address these challenges, we propose the Transposed Sparse Self-Attention (TSSA) mechanism, which improves local feature attention by restructuring the self-attention computation, without using convolutions. Additionally, we introduce a Segmented Convolutional Feed-Forward Network (SCFFN) to enhance high-frequency detail recovery and local feature acquisition, while maintaining a low parameter count. We combine TSSA, SCFFN, and a channel attention mechanism to develop TSSA-Net, an innovative network for image super-resolution. Comprehensive evaluations on classical, lightweight, and real-world SR tasks show that TSSA-Net outperforms recent methods on the Set14, Urban100, and Manga109 datasets, with improvements of 0.01–0.04 decibel (dB), 0.11–0.12 dB, and 0.01–0.08 dB, respectively. TSSA-Net achieves notable results in both metric-based and visual-based evaluations. The code is available at <span><span>https://github.com/VMC-Lab-Chen/TSSA-Net</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"153 ","pages":"Article 110823"},"PeriodicalIF":8.0000,"publicationDate":"2025-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"TSSA-Net: Transposed Sparse Self-Attention-based network for image super-resolution\",\"authors\":\"Guanhao Chen, Dan Xu, Kangjian He, Hongzhen Shi, Hao Zhang\",\"doi\":\"10.1016/j.engappai.2025.110823\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Image super-resolution (SR) aims to reconstruct high-resolution images from low-resolution ones. Transformer-based methods have recently demonstrated remarkable results, but the conventional dense self-attention mechanism fails to capture local relationships between patches, leading to suboptimal performance. Moreover, the recovery of high-frequency information, crucial for edge reconstruction, is often insufficient. To address these challenges, we propose the Transposed Sparse Self-Attention (TSSA) mechanism, which improves local feature attention by restructuring the self-attention computation, without using convolutions. Additionally, we introduce a Segmented Convolutional Feed-Forward Network (SCFFN) to enhance high-frequency detail recovery and local feature acquisition, while maintaining a low parameter count. We combine TSSA, SCFFN, and a channel attention mechanism to develop TSSA-Net, an innovative network for image super-resolution. Comprehensive evaluations on classical, lightweight, and real-world SR tasks show that TSSA-Net outperforms recent methods on the Set14, Urban100, and Manga109 datasets, with improvements of 0.01–0.04 decibel (dB), 0.11–0.12 dB, and 0.01–0.08 dB, respectively. TSSA-Net achieves notable results in both metric-based and visual-based evaluations. The code is available at <span><span>https://github.com/VMC-Lab-Chen/TSSA-Net</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":50523,\"journal\":{\"name\":\"Engineering Applications of Artificial Intelligence\",\"volume\":\"153 \",\"pages\":\"Article 110823\"},\"PeriodicalIF\":8.0000,\"publicationDate\":\"2025-04-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Engineering Applications of Artificial Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0952197625008231\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197625008231","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
TSSA-Net: Transposed Sparse Self-Attention-based network for image super-resolution
Image super-resolution (SR) aims to reconstruct high-resolution images from low-resolution ones. Transformer-based methods have recently demonstrated remarkable results, but the conventional dense self-attention mechanism fails to capture local relationships between patches, leading to suboptimal performance. Moreover, the recovery of high-frequency information, crucial for edge reconstruction, is often insufficient. To address these challenges, we propose the Transposed Sparse Self-Attention (TSSA) mechanism, which improves local feature attention by restructuring the self-attention computation, without using convolutions. Additionally, we introduce a Segmented Convolutional Feed-Forward Network (SCFFN) to enhance high-frequency detail recovery and local feature acquisition, while maintaining a low parameter count. We combine TSSA, SCFFN, and a channel attention mechanism to develop TSSA-Net, an innovative network for image super-resolution. Comprehensive evaluations on classical, lightweight, and real-world SR tasks show that TSSA-Net outperforms recent methods on the Set14, Urban100, and Manga109 datasets, with improvements of 0.01–0.04 decibel (dB), 0.11–0.12 dB, and 0.01–0.08 dB, respectively. TSSA-Net achieves notable results in both metric-based and visual-based evaluations. The code is available at https://github.com/VMC-Lab-Chen/TSSA-Net.
期刊介绍:
Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.