Efficient ultra-high-resolution hyperspectral reconstruction using a patched input spatial-spectral transformer
Author(s)
Jinhyeok An | Samsung Electronics
Wonkyung Jung | Samsung Electronics
Jintae Jang | Samsung Electronics
Jeongwook Lee | Samsung Electronics
Sung-Su Kim | Samsung Electronics
Yitae Kim | Samsung Electronics
Details
Transformers, which have demonstrated remarkable performance improvements in natural language processing, have been increasingly adopted in computer vision tasks since the introduction of the Vision Transformer (VIT). In hyperspectral image (HSI) reconstruction, Transformer-based models have gained popularity due to their ability to capture global dependencies. While these models alleviate certain limitations of convolutional neural networks (CNNs), their computational complexity scales quadratically with spatial resolution, making ultra-high-resolution reconstruction infeasible. Spectral Transformer variants have been proposed to mitigate the spatial resolution burden, yet they still face challenges in handling ultra-high-resolution imagery.
In this work, we propose a ''Patched Input Spatial-Spectral Transformer (PSST)'' that efficiently reconstructs HSIs from ultra-high-resolution RGB images. The model integrates a spatial transformer before spectral processing, enabling global context awareness while maintaining computational efficiency through in-model patch partitioning. Although performance slightly decreases for low-resolution inputs compared to state-of-the-art (SOTA) models, our method achieves the highest reconstruction quality for ultra-high-resolution inputs, outperforming existing approaches in PSNR while significantly reducing memory consumption.
Efficient ultra-high-resolution hyperspectral reconstruction using a patched input spatial-spectral transformer
Description
Date and Location: 3/2/2026 | 03:30 PM - 03:50 PM | Grand Peninsula APrimary Session Chair:
|
Session Co-Chair:
Paper Number: HVEI-210
Back to Session Gallery