{"title":"High-Generalized Unfolding Model With Coupled Spatial-Spectral Transformer for Hyperspectral Image Reconstruction","authors":"Xian-Hua Han","doi":"10.1109/TCI.2025.3564776","DOIUrl":null,"url":null,"abstract":"Deep unfolding framework has witnessed remarkable progress for hyperspectral image (HSI) reconstruction benefitting from advanced consolidation of the imaging model-driven and data-driven approaches, which are generally realized with the data reconstruction error term and the prior learning network. However, current methods still encounter challenges related to insufficient generalization and representation for the high-dimensional HSI data, manifesting in two key aspects: 1) assumption of the fixed sensing mask causing low generalization for reconstruction of the compressive measurements out of distribution; 2) imperfect prior representation network for the high-dimensional data in both spatial and spectral domains. To overcome the aforementioned issues, this study presents a high-generalized deep unfolding model using coupled spatial-spectral transformer (CS2Tr) for prior learning. Specifically, to improve the generalization capability, we synthesize the training samples with diverse masks to learn the unfolding model, and propose a mask guided-data modeling module for being incorporated with both data reconstruction term and prior learning network for degradation-aware updating and representation context modeling. To achieve robust prior representation, a coupled spatial-spectral transformer aiming at modeling both non-local spatial and spectral dependencies is introduced for capturing the 3D attributes of HSI. Moreover, we conduct the feature interaction among stages to capture rich and diverse contexts, and leverage the auxiliary losses on all stages for enhancing the recovery capability of each individual step. Extensive experiments on both simulated and real scenes have demonstrated that our proposed method outperforms the state-of-the-art HSI reconstruction approaches.","PeriodicalId":56022,"journal":{"name":"IEEE Transactions on Computational Imaging","volume":"11 ","pages":"625-637"},"PeriodicalIF":4.2000,"publicationDate":"2025-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Computational Imaging","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10978052/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Deep unfolding framework has witnessed remarkable progress for hyperspectral image (HSI) reconstruction benefitting from advanced consolidation of the imaging model-driven and data-driven approaches, which are generally realized with the data reconstruction error term and the prior learning network. However, current methods still encounter challenges related to insufficient generalization and representation for the high-dimensional HSI data, manifesting in two key aspects: 1) assumption of the fixed sensing mask causing low generalization for reconstruction of the compressive measurements out of distribution; 2) imperfect prior representation network for the high-dimensional data in both spatial and spectral domains. To overcome the aforementioned issues, this study presents a high-generalized deep unfolding model using coupled spatial-spectral transformer (CS2Tr) for prior learning. Specifically, to improve the generalization capability, we synthesize the training samples with diverse masks to learn the unfolding model, and propose a mask guided-data modeling module for being incorporated with both data reconstruction term and prior learning network for degradation-aware updating and representation context modeling. To achieve robust prior representation, a coupled spatial-spectral transformer aiming at modeling both non-local spatial and spectral dependencies is introduced for capturing the 3D attributes of HSI. Moreover, we conduct the feature interaction among stages to capture rich and diverse contexts, and leverage the auxiliary losses on all stages for enhancing the recovery capability of each individual step. Extensive experiments on both simulated and real scenes have demonstrated that our proposed method outperforms the state-of-the-art HSI reconstruction approaches.
期刊介绍:
The IEEE Transactions on Computational Imaging will publish articles where computation plays an integral role in the image formation process. Papers will cover all areas of computational imaging ranging from fundamental theoretical methods to the latest innovative computational imaging system designs. Topics of interest will include advanced algorithms and mathematical techniques, model-based data inversion, methods for image and signal recovery from sparse and incomplete data, techniques for non-traditional sensing of image data, methods for dynamic information acquisition and extraction from imaging sensors, software and hardware for efficient computation in imaging systems, and highly novel imaging system design.