TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/DiffDreamer: Towards Consistent Unsupervised Single-view S...

DiffDreamer: Towards Consistent Unsupervised Single-view Scene Extrapolation with Conditional Diffusion Models

Shengqu Cai, Eric Ryan Chan, Songyou Peng, Mohamad Shahbazi, Anton Obukhov, Luc van Gool, Gordon Wetzstein

2022-11-22ICCV 2023 1DenoisingPerpetual View Generation
PaperPDF

Abstract

Scene extrapolation -- the idea of generating novel views by flying into a given image -- is a promising, yet challenging task. For each predicted frame, a joint inpainting and 3D refinement problem has to be solved, which is ill posed and includes a high level of ambiguity. Moreover, training data for long-range scenes is difficult to obtain and usually lacks sufficient views to infer accurate camera poses. We introduce DiffDreamer, an unsupervised framework capable of synthesizing novel views depicting a long camera trajectory while training solely on internet-collected images of nature scenes. Utilizing the stochastic nature of the guided denoising steps, we train the diffusion models to refine projected RGBD images but condition the denoising steps on multiple past and future frames for inference. We demonstrate that image-conditioned diffusion models can effectively perform long-range scene extrapolation while preserving consistency significantly better than prior GAN-based methods. DiffDreamer is a powerful and efficient solution for scene extrapolation, producing impressive results despite limited supervision. Project page: https://primecai.github.io/diffdreamer.

Results

TaskDatasetMetricValueModel
Perpetual View GenerationLHQFID (first 20 steps)34.49DiffDreamer
Perpetual View GenerationLHQFID (full 100 steps)51DiffDreamer
Perpetual View GenerationLHQIS (first 20 steps)2.82DiffDreamer
Perpetual View GenerationLHQIS (full 100 steps)2.99DiffDreamer
Perpetual View GenerationLHQKID (first 20 steps)0.08DiffDreamer
Perpetual View GenerationLHQKID (full 100 steps)0.28DiffDreamer
Perpetual View GenerationLHQFID (first 20 steps)39.45InfNat-Zero
Perpetual View GenerationLHQFID (full 100 steps)26.24InfNat-Zero
Perpetual View GenerationLHQIS (first 20 steps)2.8InfNat-Zero
Perpetual View GenerationLHQIS (full 100 steps)2.72InfNat-Zero
Perpetual View GenerationLHQKID (first 20 steps)0.12InfNat-Zero
Perpetual View GenerationLHQKID (full 100 steps)0.12InfNat-Zero

Related Papers

fastWDM3D: Fast and Accurate 3D Healthy Tissue Inpainting2025-07-17Diffuman4D: 4D Consistent Human View Synthesis from Sparse-View Videos with Spatio-Temporal Diffusion Models2025-07-17Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16HUG-VAS: A Hierarchical NURBS-Based Generative Model for Aortic Geometry Synthesis and Controllable Editing2025-07-15AirLLM: Diffusion Policy-based Adaptive LoRA for Remote Fine-Tuning of LLM over the Air2025-07-15A statistical physics framework for optimal learning2025-07-10LangMamba: A Language-driven Mamba Framework for Low-dose CT Denoising with Vision-language Models2025-07-08Unconditional Diffusion for Generative Sequential Recommendation2025-07-08