TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A U-Net Based Discriminator for Generative Adversarial Net...

A U-Net Based Discriminator for Generative Adversarial Networks

Edgar Schönfeld, Bernt Schiele, Anna Khoreva

2020-02-28Data AugmentationImage GenerationConditional Image Generation
PaperPDFCode(official)CodeCode

Abstract

Among the major remaining challenges for generative adversarial networks (GANs) is the capacity to synthesize globally and locally coherent images with object shapes and textures indistinguishable from real images. To target this issue we propose an alternative U-Net based discriminator architecture, borrowing the insights from the segmentation literature. The proposed U-Net based architecture allows to provide detailed per-pixel feedback to the generator while maintaining the global coherence of synthesized images, by providing the global image feedback as well. Empowered by the per-pixel response of the discriminator, we further propose a per-pixel consistency regularization technique based on the CutMix data augmentation, encouraging the U-Net discriminator to focus more on semantic and structural changes between real and fake images. This improves the U-Net discriminator training, further enhancing the quality of generated samples. The novel discriminator improves over the state of the art in terms of the standard distribution and image quality metrics, enabling the generator to synthesize images with varying structure, appearance and levels of detail, maintaining global and local realism. Compared to the BigGAN baseline, we achieve an average improvement of 2.7 FID points across FFHQ, CelebA, and the newly introduced COCO-Animals dataset. The code is available at https://github.com/boschresearch/unetgan.

Results

TaskDatasetMetricValueModel
Image GenerationCelebA-HQ 128x128FID2.03U-Net GAN
Image GenerationCelebA-HQ 128x128Inception score3.33U-Net GAN
Image GenerationFFHQ 256 x 256FID7.48U-Net GAN
Image GenerationFFHQ 256 x 256FID11.48BigGAN
Image GenerationCelebA 128x128FID2.95U-Net GAN
Image GenerationCelebA 128x128Inception score3.43U-Net GAN
Image GenerationCOCO-AnimalsFID13.73U-Net GAN
Image GenerationCOCO-AnimalsIS12.29U-Net GAN
Image GenerationCOCO-AnimalsFID16.37BigGAN
Image GenerationCOCO-AnimalsIS11.77BigGAN
Conditional Image GenerationCOCO-AnimalsFID13.73U-Net GAN
Conditional Image GenerationCOCO-AnimalsIS12.29U-Net GAN
Conditional Image GenerationCOCO-AnimalsFID16.37BigGAN
Conditional Image GenerationCOCO-AnimalsIS11.77BigGAN

Related Papers

Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17fastWDM3D: Fast and Accurate 3D Healthy Tissue Inpainting2025-07-17Synthesizing Reality: Leveraging the Generative AI-Powered Platform Midjourney for Construction Worker Detection2025-07-17FashionPose: Text to Pose to Relight Image Generation for Personalized Fashion Visualization2025-07-17A Distributed Generative AI Approach for Heterogeneous Multi-Domain Environments under Data Sharing constraints2025-07-17Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16FADE: Adversarial Concept Erasure in Flow Models2025-07-16