TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Image-to-Image Translation with Conditional Adversarial Ne...

Image-to-Image Translation with Conditional Adversarial Networks

Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, Alexei A. Efros

2016-11-21CVPR 2017 7TranslationColorizationNuclear SegmentationCross-View Image-to-Image TranslationImage-to-Image Translation
PaperPDFCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCodeCode

Abstract

We investigate conditional adversarial networks as a general-purpose solution to image-to-image translation problems. These networks not only learn the mapping from input image to output image, but also learn a loss function to train this mapping. This makes it possible to apply the same generic approach to problems that traditionally would require very different loss formulations. We demonstrate that this approach is effective at synthesizing photos from label maps, reconstructing objects from edge maps, and colorizing images, among other tasks. Indeed, since the release of the pix2pix software associated with this paper, a large number of internet users (many of them artists) have posted their own experiments with our system, further demonstrating its wide applicability and ease of adoption without the need for parameter tweaking. As a community, we no longer hand-engineer our mapping functions, and this work suggests we can achieve reasonable results without hand-engineering our loss functions either.

Results

TaskDatasetMetricValueModel
Image-to-Image TranslationAerial-to-MapClass IOU0.26cGAN
Image-to-Image TranslationFLIRPSNR4.19pix2pix
Image-to-Image TranslationFLIRSSIM0.05pix2pix
Image-to-Image TranslationCityscapes Labels-to-PhotoClass IOU0.18pix2pix
Image-to-Image TranslationCityscapes Labels-to-PhotoPer-class Accuracy25pix2pix
Image-to-Image TranslationCityscapes Labels-to-PhotoPer-pixel Accuracy71pix2pix
Image-to-Image TranslationCityscapes Photo-to-LabelsClass IOU0.32pix2pix
Image-to-Image TranslationDayton (64x64) - ground-to-aerialSSIM0.3675Pix2pix
Image-to-Image TranslationcvusaSSIM0.3923Pix2pix
Image-to-Image TranslationDayton (64×64) - aerial-to-groundSSIM0.4808Pix2pix
Image-to-Image TranslationEgo2TopSSIM0.2213Pix2pix
Image-to-Image TranslationDayton (256×256) - ground-to-aerialSSIM0.2693Pix2pix
Image-to-Image TranslationDayton (256×256) - aerial-to-groundSSIM0.418Pix2pix
Image-to-Image TranslationFundus Fluorescein Angiogram Photographs & Colour Fundus Images of Diabetic PatientsFID48.6pix2pix
Medical Image SegmentationCell17Dice0.6351Pix2Pix
Medical Image SegmentationCell17F1-score0.6208Pix2Pix
Medical Image SegmentationCell17Hausdorff19.1441Pix2Pix
Image GenerationAerial-to-MapClass IOU0.26cGAN
Image GenerationFLIRPSNR4.19pix2pix
Image GenerationFLIRSSIM0.05pix2pix
Image GenerationCityscapes Labels-to-PhotoClass IOU0.18pix2pix
Image GenerationCityscapes Labels-to-PhotoPer-class Accuracy25pix2pix
Image GenerationCityscapes Labels-to-PhotoPer-pixel Accuracy71pix2pix
Image GenerationCityscapes Photo-to-LabelsClass IOU0.32pix2pix
Image GenerationDayton (64x64) - ground-to-aerialSSIM0.3675Pix2pix
Image GenerationcvusaSSIM0.3923Pix2pix
Image GenerationDayton (64×64) - aerial-to-groundSSIM0.4808Pix2pix
Image GenerationEgo2TopSSIM0.2213Pix2pix
Image GenerationDayton (256×256) - ground-to-aerialSSIM0.2693Pix2pix
Image GenerationDayton (256×256) - aerial-to-groundSSIM0.418Pix2pix
Image GenerationFundus Fluorescein Angiogram Photographs & Colour Fundus Images of Diabetic PatientsFID48.6pix2pix
Image ReconstructionEdge-to-HandbagsFID96.31pix2pix
Image ReconstructionEdge-to-HandbagsLPIPS0.234pix2pix
Image ReconstructionEdge-to-ShoesFID197.492pix2pix
Image ReconstructionEdge-to-ShoesLPIPS0.238pix2pix
ColorizationImageNet valFID-5K24.41cGAN
1 Image, 2*2 StitchingAerial-to-MapClass IOU0.26cGAN
1 Image, 2*2 StitchingFLIRPSNR4.19pix2pix
1 Image, 2*2 StitchingFLIRSSIM0.05pix2pix
1 Image, 2*2 StitchingCityscapes Labels-to-PhotoClass IOU0.18pix2pix
1 Image, 2*2 StitchingCityscapes Labels-to-PhotoPer-class Accuracy25pix2pix
1 Image, 2*2 StitchingCityscapes Labels-to-PhotoPer-pixel Accuracy71pix2pix
1 Image, 2*2 StitchingCityscapes Photo-to-LabelsClass IOU0.32pix2pix
1 Image, 2*2 StitchingDayton (64x64) - ground-to-aerialSSIM0.3675Pix2pix
1 Image, 2*2 StitchingcvusaSSIM0.3923Pix2pix
1 Image, 2*2 StitchingDayton (64×64) - aerial-to-groundSSIM0.4808Pix2pix
1 Image, 2*2 StitchingEgo2TopSSIM0.2213Pix2pix
1 Image, 2*2 StitchingDayton (256×256) - ground-to-aerialSSIM0.2693Pix2pix
1 Image, 2*2 StitchingDayton (256×256) - aerial-to-groundSSIM0.418Pix2pix
1 Image, 2*2 StitchingFundus Fluorescein Angiogram Photographs & Colour Fundus Images of Diabetic PatientsFID48.6pix2pix

Related Papers

A Translation of Probabilistic Event Calculus into Markov Decision Processes2025-07-17Function-to-Style Guidance of LLMs for Code Translation2025-07-15Speak2Sign3D: A Multi-modal Pipeline for English Speech to American Sign Language Animation2025-07-09Pun Intended: Multi-Agent Translation of Wordplay with Contrastive Learning and Phonetic-Semantic Embeddings2025-07-09Unconditional Diffusion for Generative Sequential Recommendation2025-07-08GRAFT: A Graph-based Flow-aware Agentic Framework for Document-level Machine Translation2025-07-04TransLaw: Benchmarking Large Language Models in Multi-Agent Simulation of the Collaborative Translation2025-07-01CycleVAR: Repurposing Autoregressive Model for Unsupervised One-Step Image Translation2025-06-29