TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/ASteISR: Adapting Single Image Super-resolution Pre-traine...

ASteISR: Adapting Single Image Super-resolution Pre-trained Model for Efficient Stereo Image Super-resolution

Yuanbo Zhou, Yuyang Xue, Wei Deng, Xinlin Zhang, Qinquan Gao, Tong Tong

2024-07-04Super-ResolutionImage Super-Resolutionparameter-efficient fine-tuningStereo Image Super-Resolution
PaperPDFCode(official)

Abstract

Despite advances in the paradigm of pre-training then fine-tuning in low-level vision tasks, significant challenges persist particularly regarding the increased size of pre-trained models such as memory usage and training time. Another concern often encountered is the unsatisfying results yielded when directly applying pre-trained single-image models to multi-image domain. In this paper, we propose a efficient method for transferring a pre-trained single-image super-resolution (SISR) transformer network to the domain of stereo image super-resolution (SteISR) through a parameter-efficient fine-tuning (PEFT) method. Specifically, we introduce the concept of stereo adapters and spatial adapters which are incorporated into the pre-trained SISR transformer network. Subsequently, the pre-trained SISR model is frozen, enabling us to fine-tune the adapters using stereo datasets along. By adopting this training method, we enhance the ability of the SISR model to accurately infer stereo images by 0.79dB on the Flickr1024 dataset. This method allows us to train only 4.8% of the original model parameters, achieving state-of-the-art performance on four commonly used SteISR benchmarks. Compared to the more complicated full fine-tuning approach, our method reduces training time and memory consumption by 57% and 15%, respectively.

Results

TaskDatasetMetricValueModel
Super-ResolutionMiddlebury - 2x upscalingPSNR36.6ASteISR
Super-ResolutionFlickr1024 - 2x upscalingPSNR30.33ASteISR
Super-ResolutionKITTI2012 - 2x upscalingPSNR31.86ASteISR
Super-ResolutionKITTI2015 - 2x upscalingPSNR31.48ASteISR
Image Super-ResolutionMiddlebury - 2x upscalingPSNR36.6ASteISR
Image Super-ResolutionFlickr1024 - 2x upscalingPSNR30.33ASteISR
Image Super-ResolutionKITTI2012 - 2x upscalingPSNR31.86ASteISR
Image Super-ResolutionKITTI2015 - 2x upscalingPSNR31.48ASteISR
3D Object Super-ResolutionMiddlebury - 2x upscalingPSNR36.6ASteISR
3D Object Super-ResolutionFlickr1024 - 2x upscalingPSNR30.33ASteISR
3D Object Super-ResolutionKITTI2012 - 2x upscalingPSNR31.86ASteISR
3D Object Super-ResolutionKITTI2015 - 2x upscalingPSNR31.48ASteISR
16kMiddlebury - 2x upscalingPSNR36.6ASteISR
16kFlickr1024 - 2x upscalingPSNR30.33ASteISR
16kKITTI2012 - 2x upscalingPSNR31.86ASteISR
16kKITTI2015 - 2x upscalingPSNR31.48ASteISR

Related Papers

SpectraLift: Physics-Guided Spectral-Inversion Network for Self-Supervised Hyperspectral Image Super-Resolution2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17IM-LUT: Interpolation Mixing Look-Up Tables for Image Super-Resolution2025-07-14PanoDiff-SR: Synthesizing Dental Panoramic Radiographs using Diffusion and Super-resolution2025-07-12HNOSeg-XS: Extremely Small Hartley Neural Operator for Efficient and Resolution-Robust 3D Image Segmentation2025-07-104KAgent: Agentic Any Image to 4K Super-Resolution2025-07-09LoSiA: Efficient High-Rank Fine-Tuning via Subnet Localization and Optimization2025-07-06EAMamba: Efficient All-Around Vision State Space Model for Image Restoration2025-06-27