TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Learning Spatiotemporal Frequency-Transformer for Compress...

Learning Spatiotemporal Frequency-Transformer for Compressed Video Super-Resolution

Zhongwei Qiu, Huan Yang, Jianlong Fu, Dongmei Fu

2022-08-05Super-ResolutionVideo Super-ResolutionVideo Enhancement
PaperPDFCode(official)

Abstract

Compressed video super-resolution (VSR) aims to restore high-resolution frames from compressed low-resolution counterparts. Most recent VSR approaches often enhance an input frame by borrowing relevant textures from neighboring video frames. Although some progress has been made, there are grand challenges to effectively extract and transfer high-quality textures from compressed videos where most frames are usually highly degraded. In this paper, we propose a novel Frequency-Transformer for compressed video super-resolution (FTVSR) that conducts self-attention over a joint space-time-frequency domain. First, we divide a video frame into patches, and transform each patch into DCT spectral maps in which each channel represents a frequency band. Such a design enables a fine-grained level self-attention on each frequency band, so that real visual texture can be distinguished from artifacts, and further utilized for video frame restoration. Second, we study different self-attention schemes, and discover that a divided attention which conducts a joint space-frequency attention before applying temporal attention on each frequency band, leads to the best video enhancement quality. Experimental results on two widely-used video super-resolution benchmarks show that FTVSR outperforms state-of-the-art approaches on both uncompressed and compressed videos with clear visual margins. Code is available at https://github.com/researchmm/FTVSR.

Results

TaskDatasetMetricValueModel
Super-ResolutionREDS4- 4x upscalingPSNR31.82FTVSR
Super-ResolutionREDS4- 4x upscalingSSIM0.896FTVSR
3D Human Pose EstimationREDS4- 4x upscalingPSNR31.82FTVSR
3D Human Pose EstimationREDS4- 4x upscalingSSIM0.896FTVSR
VideoREDS4- 4x upscalingPSNR31.82FTVSR
VideoREDS4- 4x upscalingSSIM0.896FTVSR
Pose EstimationREDS4- 4x upscalingPSNR31.82FTVSR
Pose EstimationREDS4- 4x upscalingSSIM0.896FTVSR
3DREDS4- 4x upscalingPSNR31.82FTVSR
3DREDS4- 4x upscalingSSIM0.896FTVSR
3D Face AnimationREDS4- 4x upscalingPSNR31.82FTVSR
3D Face AnimationREDS4- 4x upscalingSSIM0.896FTVSR
2D Human Pose EstimationREDS4- 4x upscalingPSNR31.82FTVSR
2D Human Pose EstimationREDS4- 4x upscalingSSIM0.896FTVSR
3D Absolute Human Pose EstimationREDS4- 4x upscalingPSNR31.82FTVSR
3D Absolute Human Pose EstimationREDS4- 4x upscalingSSIM0.896FTVSR
Video Super-ResolutionREDS4- 4x upscalingPSNR31.82FTVSR
Video Super-ResolutionREDS4- 4x upscalingSSIM0.896FTVSR
3D Object Super-ResolutionREDS4- 4x upscalingPSNR31.82FTVSR
3D Object Super-ResolutionREDS4- 4x upscalingSSIM0.896FTVSR
1 Image, 2*2 StitchiREDS4- 4x upscalingPSNR31.82FTVSR
1 Image, 2*2 StitchiREDS4- 4x upscalingSSIM0.896FTVSR

Related Papers

SpectraLift: Physics-Guided Spectral-Inversion Network for Self-Supervised Hyperspectral Image Super-Resolution2025-07-17IM-LUT: Interpolation Mixing Look-Up Tables for Image Super-Resolution2025-07-14PanoDiff-SR: Synthesizing Dental Panoramic Radiographs using Diffusion and Super-resolution2025-07-12HNOSeg-XS: Extremely Small Hartley Neural Operator for Efficient and Resolution-Robust 3D Image Segmentation2025-07-104KAgent: Agentic Any Image to 4K Super-Resolution2025-07-09EAMamba: Efficient All-Around Vision State Space Model for Image Restoration2025-06-27Leveraging Vision-Language Models to Select Trustworthy Super-Resolution Samples Generated by Diffusion Models2025-06-25Unsupervised Image Super-Resolution Reconstruction Based on Real-World Degradation Patterns2025-06-20