TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Unified Continuous Generative Models

Unified Continuous Generative Models

Peng Sun, Yi Jiang, Tao Lin

2025-05-12Image Generation
PaperPDFCode(official)

Abstract

Recent advances in continuous generative models, including multi-step approaches like diffusion and flow-matching (typically requiring 8-1000 sampling steps) and few-step methods such as consistency models (typically 1-8 steps), have demonstrated impressive generative performance. However, existing work often treats these approaches as distinct paradigms, resulting in separate training and sampling methodologies. We introduce a unified framework for training, sampling, and analyzing these models. Our implementation, the Unified Continuous Generative Models Trainer and Sampler (UCGM-{T,S}), achieves state-of-the-art (SOTA) performance. For example, on ImageNet 256x256 using a 675M diffusion transformer, UCGM-T trains a multi-step model achieving 1.30 FID in 20 steps and a few-step model reaching 1.42 FID in just 2 steps. Additionally, applying UCGM-S to a pre-trained model (previously 1.26 FID at 250 steps) improves performance to 1.06 FID in only 40 steps. Code is available at: https://github.com/LINs-lab/UCGM.

Results

TaskDatasetMetricValueModel
Image GenerationImageNet 512x512FID1.24DDT-XL/2 + UCGM-S (SD-VAE + 150 sampling steps + CFG)
Image GenerationImageNet 512x512NFE300DDT-XL/2 + UCGM-S (SD-VAE + 150 sampling steps + CFG)
Image GenerationImageNet 512x512FID1.25DDT-XL/2 + UCGM-S (SD-VAE + 100 sampling steps + CFG)
Image GenerationImageNet 512x512NFE200DDT-XL/2 + UCGM-S (SD-VAE + 100 sampling steps + CFG)
Image GenerationImageNet 256x256FID1.06SiT-XL/2 + UCGM-S (E2E-VAE + 40 sampling steps + CFG)
Image GenerationImageNet 256x256NFE80SiT-XL/2 + UCGM-S (E2E-VAE + 40 sampling steps + CFG)
Image GenerationImageNet 256x256FID1.21UCGM-XL/2 (VA-VAE + 30 sampling steps, without guidance)
Image GenerationImageNet 256x256NFE30UCGM-XL/2 (VA-VAE + 30 sampling steps, without guidance)
Image GenerationImageNet 256x256FID1.21UCGM-XL/2 (E2E-VAE + 40 sampling steps, without guidance)
Image GenerationImageNet 256x256NFE40UCGM-XL/2 (E2E-VAE + 40 sampling steps, without guidance)
Image GenerationImageNet 256x256FID1.21LightningDiT + UCGM-S (VA-VAE + 50 sampling steps + CFG)
Image GenerationImageNet 256x256NFE100LightningDiT + UCGM-S (VA-VAE + 50 sampling steps + CFG)

Related Papers

fastWDM3D: Fast and Accurate 3D Healthy Tissue Inpainting2025-07-17Synthesizing Reality: Leveraging the Generative AI-Powered Platform Midjourney for Construction Worker Detection2025-07-17FashionPose: Text to Pose to Relight Image Generation for Personalized Fashion Visualization2025-07-17A Distributed Generative AI Approach for Heterogeneous Multi-Domain Environments under Data Sharing constraints2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17FADE: Adversarial Concept Erasure in Flow Models2025-07-16CharaConsist: Fine-Grained Consistent Character Generation2025-07-15CATVis: Context-Aware Thought Visualization2025-07-15