TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/GRAF: Generative Radiance Fields for 3D-Aware Image Synthe...

GRAF: Generative Radiance Fields for 3D-Aware Image Synthesis

Katja Schwarz, Yiyi Liao, Michael Niemeyer, Andreas Geiger

2020-07-05NeurIPS 2020 12Novel View SynthesisScene Generation3D-Aware Image SynthesisImage Generation
PaperPDFCode(official)

Abstract

While 2D generative adversarial networks have enabled high-resolution image synthesis, they largely lack an understanding of the 3D world and the image formation process. Thus, they do not provide precise control over camera viewpoint or object pose. To address this problem, several recent approaches leverage intermediate voxel-based representations in combination with differentiable rendering. However, existing methods either produce low image resolution or fall short in disentangling camera and scene properties, e.g., the object identity may vary with the viewpoint. In this paper, we propose a generative model for radiance fields which have recently proven successful for novel view synthesis of a single scene. In contrast to voxel-based representations, radiance fields are not confined to a coarse discretization of the 3D space, yet allow for disentangling camera and scene properties while degrading gracefully in the presence of reconstruction ambiguity. By introducing a multi-scale patch-based discriminator, we demonstrate synthesis of high-resolution images while training our model from unposed 2D images alone. We systematically analyze our approach on several challenging synthetic and real-world datasets. Our experiments reveal that radiance fields are a powerful representation for generative image synthesis, leading to 3D consistent models that render with high fidelity.

Results

TaskDatasetMetricValueModel
Image GenerationVizDoomFID47.5GRAF
Image GenerationVizDoomFID (SwAV)5.44GRAF
Image GenerationReplicaFID65.37GRAF
Image GenerationReplicaFID (SwAV)5.76GRAF
Scene GenerationReplicaFID65.37GRAF
Scene GenerationReplicaSwAV-FID5.76GRAF
Scene GenerationAVDFID62.59GRAF
Scene GenerationAVDSwAV-FID6.95GRAF
Scene GenerationVizDoomFID47.5GRAF
Scene GenerationVizDoomSwAV-FID5.44GRAF
16kReplicaFID65.37GRAF
16kReplicaSwAV-FID5.76GRAF
16kAVDFID62.59GRAF
16kAVDSwAV-FID6.95GRAF
16kVizDoomFID47.5GRAF
16kVizDoomSwAV-FID5.44GRAF

Related Papers

World Model-Based End-to-End Scene Generation for Accident Anticipation in Autonomous Driving2025-07-17fastWDM3D: Fast and Accurate 3D Healthy Tissue Inpainting2025-07-17Synthesizing Reality: Leveraging the Generative AI-Powered Platform Midjourney for Construction Worker Detection2025-07-17FashionPose: Text to Pose to Relight Image Generation for Personalized Fashion Visualization2025-07-17A Distributed Generative AI Approach for Heterogeneous Multi-Domain Environments under Data Sharing constraints2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17FADE: Adversarial Concept Erasure in Flow Models2025-07-16Physically Based Neural LiDAR Resimulation2025-07-15