TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Blur-aware Spatio-temporal Sparse Transformer for Video De...

Blur-aware Spatio-temporal Sparse Transformer for Video Deblurring

Huicong Zhang, Haozhe Xie, Hongxun Yao

2024-06-11CVPR 2024 1DeblurringOptical Flow EstimationVideo Deblurring
PaperPDFCode(official)

Abstract

Video deblurring relies on leveraging information from other frames in the video sequence to restore the blurred regions in the current frame. Mainstream approaches employ bidirectional feature propagation, spatio-temporal transformers, or a combination of both to extract information from the video sequence. However, limitations in memory and computational resources constraints the temporal window length of the spatio-temporal transformer, preventing the extraction of longer temporal contextual information from the video sequence. Additionally, bidirectional feature propagation is highly sensitive to inaccurate optical flow in blurry frames, leading to error accumulation during the propagation process. To address these issues, we propose \textbf{BSSTNet}, \textbf{B}lur-aware \textbf{S}patio-temporal \textbf{S}parse \textbf{T}ransformer Network. It introduces the blur map, which converts the originally dense attention into a sparse form, enabling a more extensive utilization of information throughout the entire video sequence. Specifically, BSSTNet (1) uses a longer temporal window in the transformer, leveraging information from more distant frames to restore the blurry pixels in the current frame. (2) introduces bidirectional feature propagation guided by blur maps, which reduces error accumulation caused by the blur frame. The experimental results demonstrate the proposed BSSTNet outperforms the state-of-the-art methods on the GoPro and DVD datasets.

Results

TaskDatasetMetricValueModel
DeblurringDVDPSNR34.95BSSTNet
DeblurringDVDSSIM0.9703BSSTNet
DeblurringGoProPSNR35.98BSSTNet
DeblurringGoProSSIM0.9792BSSTNet
2D ClassificationDVDPSNR34.95BSSTNet
2D ClassificationDVDSSIM0.9703BSSTNet
2D ClassificationGoProPSNR35.98BSSTNet
2D ClassificationGoProSSIM0.9792BSSTNet
10-shot image generationDVDPSNR34.95BSSTNet
10-shot image generationDVDSSIM0.9703BSSTNet
10-shot image generationGoProPSNR35.98BSSTNet
10-shot image generationGoProSSIM0.9792BSSTNet
Blind Image DeblurringDVDPSNR34.95BSSTNet
Blind Image DeblurringDVDSSIM0.9703BSSTNet
Blind Image DeblurringGoProPSNR35.98BSSTNet
Blind Image DeblurringGoProSSIM0.9792BSSTNet

Related Papers

Channel-wise Motion Features for Efficient Motion Segmentation2025-07-17Generative Latent Kernel Modeling for Blind Motion Deblurring2025-07-12An Efficient Approach for Muscle Segmentation and 3D Reconstruction Using Keypoint Tracking in MRI Scan2025-07-11Learning to Track Any Points from Human Motion2025-07-08TLB-VFI: Temporal-Aware Latent Brownian Bridge Diffusion for Video Frame Interpolation2025-07-07MEMFOF: High-Resolution Training for Memory-Efficient Multi-Frame Optical Flow Estimation2025-06-29EAMamba: Efficient All-Around Vision State Space Model for Image Restoration2025-06-27EndoFlow-SLAM: Real-Time Endoscopic SLAM with Flow-Constrained Gaussian Splatting2025-06-26