TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/GaussianCity: Generative Gaussian Splatting for Unbounded ...

GaussianCity: Generative Gaussian Splatting for Unbounded 3D City Generation

Haozhe Xie, Zhaoxi Chen, Fangzhou Hong, Ziwei Liu

2024-06-103D GenerationScene Generation
PaperPDFCode(official)

Abstract

3D city generation with NeRF-based methods shows promising generation results but is computationally inefficient. Recently 3D Gaussian Splatting (3D-GS) has emerged as a highly efficient alternative for object-level 3D generation. However, adapting 3D-GS from finite-scale 3D objects and humans to infinite-scale 3D cities is non-trivial. Unbounded 3D city generation entails significant storage overhead (out-of-memory issues), arising from the need to expand points to billions, often demanding hundreds of Gigabytes of VRAM for a city scene spanning 10km^2. In this paper, we propose GaussianCity, a generative Gaussian Splatting framework dedicated to efficiently synthesizing unbounded 3D cities with a single feed-forward pass. Our key insights are two-fold: 1) Compact 3D Scene Representation: We introduce BEV-Point as a highly compact intermediate representation, ensuring that the growth in VRAM usage for unbounded scenes remains constant, thus enabling unbounded city generation. 2) Spatial-aware Gaussian Attribute Decoder: We present spatial-aware BEV-Point decoder to produce 3D Gaussian attributes, which leverages Point Serializer to integrate the structural and contextual characteristics of BEV points. Extensive experiments demonstrate that GaussianCity achieves state-of-the-art results in both drone-view and street-view 3D city generation. Notably, compared to CityDreamer, GaussianCity exhibits superior performance with a speedup of 60 times (10.72 FPS v.s. 0.18 FPS).

Results

TaskDatasetMetricValueModel
Scene GenerationKITTIFID29.5GaussianCity
Scene GenerationKITTIKID0.017GaussianCity
Scene GenerationGoogleEarthCamera Error0.057GaussianCity
Scene GenerationGoogleEarthDepth Error0.136GaussianCity
Scene GenerationGoogleEarthFID86.94GaussianCity
Scene GenerationGoogleEarthKID0.09GaussianCity
16kKITTIFID29.5GaussianCity
16kKITTIKID0.017GaussianCity
16kGoogleEarthCamera Error0.057GaussianCity
16kGoogleEarthDepth Error0.136GaussianCity
16kGoogleEarthFID86.94GaussianCity
16kGoogleEarthKID0.09GaussianCity

Related Papers

AutoPartGen: Autogressive 3D Part Generation and Discovery2025-07-17World Model-Based End-to-End Scene Generation for Accident Anticipation in Autonomous Driving2025-07-17PhysX: Physical-Grounded 3D Asset Generation2025-07-16$I^{2}$-World: Intra-Inter Tokenization for Efficient Dynamic 4D Scene Forecasting2025-07-12DreamArt: Generating Interactable Articulated Objects from a Single Image2025-07-08OmniPart: Part-Aware 3D Generation with Semantic Decoupling and Structural Cohesion2025-07-08Acquiring and Adapting Priors for Novel Tasks via Neural Meta-Architectures2025-07-07Voyaging into Unbounded Dynamic Scenes from a Single View2025-07-05