TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Curriculum Direct Preference Optimization for Diffusion an...

Curriculum Direct Preference Optimization for Diffusion and Consistency Models

Florinel-Alin Croitoru, Vlad Hondru, Radu Tudor Ionescu, Nicu Sebe, Mubarak Shah

2024-05-22CVPR 2025 1Text-to-Image GenerationText to Image GenerationImage Generation
PaperPDFCode(official)

Abstract

Direct Preference Optimization (DPO) has been proposed as an effective and efficient alternative to reinforcement learning from human feedback (RLHF). In this paper, we propose a novel and enhanced version of DPO based on curriculum learning for text-to-image generation. Our method is divided into two training stages. First, a ranking of the examples generated for each prompt is obtained by employing a reward model. Then, increasingly difficult pairs of examples are sampled and provided to a text-to-image generative (diffusion or consistency) model. Generated samples that are far apart in the ranking are considered to form easy pairs, while those that are close in the ranking form hard pairs. In other words, we use the rank difference between samples as a measure of difficulty. The sampled pairs are split into batches according to their difficulty levels, which are gradually used to train the generative model. Our approach, Curriculum DPO, is compared against state-of-the-art fine-tuning approaches on nine benchmarks, outperforming the competing methods in terms of text alignment, aesthetics and human preference. Our code is available at https://github.com/CroitoruAlin/Curriculum-DPO.

Results

TaskDatasetMetricValueModel
Image GenerationDrawBenchAesthetics (Laion Aesthtetics Predictor)6.1829LCM (Curriculum DPO)
Image GenerationDrawBenchHuman Preference Alignement (HPSv2)0.2851LCM (Curriculum DPO)
Image GenerationDrawBenchText Alignement (SentenceBERT)0.5812LCM (Curriculum DPO)
Image GenerationDrawBenchAesthetics (Laion Aesthtetics Predictor)5.706Stable Diffusion 1.5 (Curriculum DPO)
Image GenerationDrawBenchHuman Preference Alignement (HPSv2)0.2681Stable Diffusion 1.5 (Curriculum DPO)
Image GenerationDrawBenchText Alignement (SentenceBERT)0.6234Stable Diffusion 1.5 (Curriculum DPO)
Text-to-Image GenerationDrawBenchAesthetics (Laion Aesthtetics Predictor)6.1829LCM (Curriculum DPO)
Text-to-Image GenerationDrawBenchHuman Preference Alignement (HPSv2)0.2851LCM (Curriculum DPO)
Text-to-Image GenerationDrawBenchText Alignement (SentenceBERT)0.5812LCM (Curriculum DPO)
Text-to-Image GenerationDrawBenchAesthetics (Laion Aesthtetics Predictor)5.706Stable Diffusion 1.5 (Curriculum DPO)
Text-to-Image GenerationDrawBenchHuman Preference Alignement (HPSv2)0.2681Stable Diffusion 1.5 (Curriculum DPO)
Text-to-Image GenerationDrawBenchText Alignement (SentenceBERT)0.6234Stable Diffusion 1.5 (Curriculum DPO)
10-shot image generationDrawBenchAesthetics (Laion Aesthtetics Predictor)6.1829LCM (Curriculum DPO)
10-shot image generationDrawBenchHuman Preference Alignement (HPSv2)0.2851LCM (Curriculum DPO)
10-shot image generationDrawBenchText Alignement (SentenceBERT)0.5812LCM (Curriculum DPO)
10-shot image generationDrawBenchAesthetics (Laion Aesthtetics Predictor)5.706Stable Diffusion 1.5 (Curriculum DPO)
10-shot image generationDrawBenchHuman Preference Alignement (HPSv2)0.2681Stable Diffusion 1.5 (Curriculum DPO)
10-shot image generationDrawBenchText Alignement (SentenceBERT)0.6234Stable Diffusion 1.5 (Curriculum DPO)
1 Image, 2*2 StitchiDrawBenchAesthetics (Laion Aesthtetics Predictor)6.1829LCM (Curriculum DPO)
1 Image, 2*2 StitchiDrawBenchHuman Preference Alignement (HPSv2)0.2851LCM (Curriculum DPO)
1 Image, 2*2 StitchiDrawBenchText Alignement (SentenceBERT)0.5812LCM (Curriculum DPO)
1 Image, 2*2 StitchiDrawBenchAesthetics (Laion Aesthtetics Predictor)5.706Stable Diffusion 1.5 (Curriculum DPO)
1 Image, 2*2 StitchiDrawBenchHuman Preference Alignement (HPSv2)0.2681Stable Diffusion 1.5 (Curriculum DPO)
1 Image, 2*2 StitchiDrawBenchText Alignement (SentenceBERT)0.6234Stable Diffusion 1.5 (Curriculum DPO)

Related Papers

fastWDM3D: Fast and Accurate 3D Healthy Tissue Inpainting2025-07-17Synthesizing Reality: Leveraging the Generative AI-Powered Platform Midjourney for Construction Worker Detection2025-07-17FashionPose: Text to Pose to Relight Image Generation for Personalized Fashion Visualization2025-07-17A Distributed Generative AI Approach for Heterogeneous Multi-Domain Environments under Data Sharing constraints2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17FADE: Adversarial Concept Erasure in Flow Models2025-07-16CharaConsist: Fine-Grained Consistent Character Generation2025-07-15CATVis: Context-Aware Thought Visualization2025-07-15