TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Fantasia3D: Disentangling Geometry and Appearance for High...

Fantasia3D: Disentangling Geometry and Appearance for High-quality Text-to-3D Content Creation

Rui Chen, Yongwei Chen, Ningxin Jiao, Kui Jia

2023-03-24ICCV 2023 1Text to 3D
PaperPDFCode(official)CodeCode

Abstract

Automatic 3D content creation has achieved rapid progress recently due to the availability of pre-trained, large language models and image diffusion models, forming the emerging topic of text-to-3D content creation. Existing text-to-3D methods commonly use implicit scene representations, which couple the geometry and appearance via volume rendering and are suboptimal in terms of recovering finer geometries and achieving photorealistic rendering; consequently, they are less effective for generating high-quality 3D assets. In this work, we propose a new method of Fantasia3D for high-quality text-to-3D content creation. Key to Fantasia3D is the disentangled modeling and learning of geometry and appearance. For geometry learning, we rely on a hybrid scene representation, and propose to encode surface normal extracted from the representation as the input of the image diffusion model. For appearance modeling, we introduce the spatially varying bidirectional reflectance distribution function (BRDF) into the text-to-3D task, and learn the surface material for photorealistic rendering of the generated surface. Our disentangled framework is more compatible with popular graphics engines, supporting relighting, editing, and physical simulation of the generated 3D assets. We conduct thorough experiments that show the advantages of our method over existing ones under different text-to-3D task settings. Project page and source codes: https://fantasia3d.github.io/.

Results

TaskDatasetMetricValueModel
3DT$^3$BenchAvg24Fantasia3D
Text to Image GenerationT$^3$BenchAvg24Fantasia3D
Text to 3DT$^3$BenchAvg24Fantasia3D

Related Papers

Acquiring and Adapting Priors for Novel Tasks via Neural Meta-Architectures2025-07-07DreamAnywhere: Object-Centric Panoramic 3D Scene Generation2025-06-25Dive3D: Diverse Distillation-based Text-to-3D Generation via Score Implicit Matching2025-06-16EmbodiedGen: Towards a Generative 3D World Engine for Embodied Intelligence2025-06-12DreamCS: Geometry-Aware Text-to-3D Generation with Unpaired 3D Reward Supervision2025-06-11R3D2: Realistic 3D Asset Insertion via Diffusion for Autonomous Driving Simulation2025-06-09AI-powered Contextual 3D Environment Generation: A Systematic Review2025-06-05ArtiScene: Language-Driven Artistic 3D Scene Generation Through Image Intermediary2025-05-31