TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Mask-guided Spectral-wise Transformer for Efficient Hypers...

Mask-guided Spectral-wise Transformer for Efficient Hyperspectral Image Reconstruction

Yuanhao Cai, Jing Lin, Xiaowan Hu, Haoqian Wang, Xin Yuan, Yulun Zhang, Radu Timofte, Luc van Gool

2021-11-15CVPR 2022 1Spectral ReconstructionImage ReconstructionCompressive SensingImage Restoration
PaperPDFCodeCodeCodeCode(official)

Abstract

Hyperspectral image (HSI) reconstruction aims to recover the 3D spatial-spectral signal from a 2D measurement in the coded aperture snapshot spectral imaging (CASSI) system. The HSI representations are highly similar and correlated across the spectral dimension. Modeling the inter-spectra interactions is beneficial for HSI reconstruction. However, existing CNN-based methods show limitations in capturing spectral-wise similarity and long-range dependencies. Besides, the HSI information is modulated by a coded aperture (physical mask) in CASSI. Nonetheless, current algorithms have not fully explored the guidance effect of the mask for HSI restoration. In this paper, we propose a novel framework, Mask-guided Spectral-wise Transformer (MST), for HSI reconstruction. Specifically, we present a Spectral-wise Multi-head Self-Attention (S-MSA) that treats each spectral feature as a token and calculates self-attention along the spectral dimension. In addition, we customize a Mask-guided Mechanism (MM) that directs S-MSA to pay attention to spatial regions with high-fidelity spectral representations. Extensive experiments show that our MST significantly outperforms state-of-the-art (SOTA) methods on simulation and real HSI datasets while requiring dramatically cheaper computational and memory costs. Code and pre-trained models are available at https://github.com/caiyuanhao1998/MST/

Results

TaskDatasetMetricValueModel
Image RestorationKAISTPSNR35.18MST-L
Image RestorationKAISTSSIM0.948MST-L
Image RestorationCAVEPSNR35.18MST-L
Image RestorationCAVESSIM0.948MST-L
Image RestorationReal HSIUser Study Score12MST
Image RestorationARAD-1KMRAE0.1772MST-L
Image RestorationARAD-1KPSNR33.9MST-L
Image RestorationARAD-1KRMSE0.0256MST-L
10-shot image generationKAISTPSNR35.18MST-L
10-shot image generationKAISTSSIM0.948MST-L
10-shot image generationCAVEPSNR35.18MST-L
10-shot image generationCAVESSIM0.948MST-L
10-shot image generationReal HSIUser Study Score12MST
10-shot image generationARAD-1KMRAE0.1772MST-L
10-shot image generationARAD-1KPSNR33.9MST-L
10-shot image generationARAD-1KRMSE0.0256MST-L

Related Papers

SpectraLift: Physics-Guided Spectral-Inversion Network for Self-Supervised Hyperspectral Image Super-Resolution2025-07-17Disentangling coincident cell events using deep transfer learning and compressive sensing2025-07-17Unsupervised Part Discovery via Descriptor-Based Masked Image Restoration with Optimized Constraints2025-07-16The model is the message: Lightweight convolutional autoencoders applied to noisy imaging data for planetary science and astrobiology2025-07-153D Magnetic Inverse Routine for Single-Segment Magnetic Field Images2025-07-15MGVQ: Could VQ-VAE Beat VAE? A Generalizable Tokenizer with Multi-group Quantization2025-07-14Vision Foundation Models as Effective Visual Tokenizers for Autoregressive Image Generation2025-07-11LangMamba: A Language-driven Mamba Framework for Low-dose CT Denoising with Vision-language Models2025-07-08