TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Towards An End-to-End Framework for Flow-Guided Video Inpa...

Towards An End-to-End Framework for Flow-Guided Video Inpainting

Zhen Li, Cheng-Ze Lu, Jianhua Qin, Chun-Le Guo, Ming-Ming Cheng

2022-04-06CVPR 2022 1Optical Flow EstimationSeeing Beyond the VisibleHallucinationVideo Inpainting
PaperPDFCode(official)Code

Abstract

Optical flow, which captures motion information across frames, is exploited in recent video inpainting methods through propagating pixels along its trajectories. However, the hand-crafted flow-based processes in these methods are applied separately to form the whole inpainting pipeline. Thus, these methods are less efficient and rely heavily on the intermediate results from earlier stages. In this paper, we propose an End-to-End framework for Flow-Guided Video Inpainting (E$^2$FGVI) through elaborately designed three trainable modules, namely, flow completion, feature propagation, and content hallucination modules. The three modules correspond with the three stages of previous flow-based methods but can be jointly optimized, leading to a more efficient and effective inpainting process. Experimental results demonstrate that the proposed method outperforms state-of-the-art methods both qualitatively and quantitatively and shows promising efficiency. The code is available at https://github.com/MCG-NKU/E2FGVI.

Results

TaskDatasetMetricValueModel
3DDAVISEwarp0.1315E2FGVI
3DDAVISPSNR33.01E2FGVI
3DDAVISSSIM0.9721E2FGVI
3DDAVISVFID0.116E2FGVI
3DYouTube-VOS 2018Ewarp0.0864E2FGVI
3DYouTube-VOS 2018PSNR33.71E2FGVI
3DYouTube-VOS 2018SSIM0.97E2FGVI
3DYouTube-VOS 2018VFID0.046E2FGVI
3DHQVI (240p)LPIPS0.0401E2FGVI
3DHQVI (240p)PSNR30.63E2FGVI
3DHQVI (240p)SSIM0.9427E2FGVI
3DHQVI (240p)VFID0.1885E2FGVI
Video InpaintingDAVISEwarp0.1315E2FGVI
Video InpaintingDAVISPSNR33.01E2FGVI
Video InpaintingDAVISSSIM0.9721E2FGVI
Video InpaintingDAVISVFID0.116E2FGVI
Video InpaintingYouTube-VOS 2018Ewarp0.0864E2FGVI
Video InpaintingYouTube-VOS 2018PSNR33.71E2FGVI
Video InpaintingYouTube-VOS 2018SSIM0.97E2FGVI
Video InpaintingYouTube-VOS 2018VFID0.046E2FGVI
Video InpaintingHQVI (240p)LPIPS0.0401E2FGVI
Video InpaintingHQVI (240p)PSNR30.63E2FGVI
Video InpaintingHQVI (240p)SSIM0.9427E2FGVI
Video InpaintingHQVI (240p)VFID0.1885E2FGVI
Seeing Beyond the VisibleKITTI360-EXAverage PSNR19.45E2FGVI

Related Papers

Channel-wise Motion Features for Efficient Motion Segmentation2025-07-17Mitigating Object Hallucinations via Sentence-Level Early Intervention2025-07-16An Efficient Approach for Muscle Segmentation and 3D Reconstruction Using Keypoint Tracking in MRI Scan2025-07-11ByDeWay: Boost Your multimodal LLM with DEpth prompting in a Training-Free Way2025-07-11Learning to Track Any Points from Human Motion2025-07-08UQLM: A Python Package for Uncertainty Quantification in Large Language Models2025-07-08TLB-VFI: Temporal-Aware Latent Brownian Bridge Diffusion for Video Frame Interpolation2025-07-07DeepRetro: Retrosynthetic Pathway Discovery using Iterative LLM Reasoning2025-07-07