TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/SESAME: Semantic Editing of Scenes by Adding, Manipulating...

SESAME: Semantic Editing of Scenes by Adding, Manipulating or Erasing Objects

Evangelos Ntavelis, Andrés Romero, Iason Kastanis, Luc van Gool, Radu Timofte

2020-04-10ECCV 2020 8Image GenerationImage ManipulationImage-to-Image Translation
PaperPDFCode(official)

Abstract

Recent advances in image generation gave rise to powerful tools for semantic image editing. However, existing approaches can either operate on a single image or require an abundance of additional information. They are not capable of handling the complete set of editing operations, that is addition, manipulation or removal of semantic concepts. To address these limitations, we propose SESAME, a novel generator-discriminator pair for Semantic Editing of Scenes by Adding, Manipulating or Erasing objects. In our setup, the user provides the semantic labels of the areas to be edited and the generator synthesizes the corresponding pixels. In contrast to previous methods that employ a discriminator that trivially concatenates semantics and image as an input, the SESAME discriminator is composed of two input streams that independently process the image and its semantics, using the latter to manipulate the results of the former. We evaluate our model on a diverse set of datasets and report state-of-the-art performance on two tasks: (a) image manipulation and (b) image generation conditioned on semantic labels.

Results

TaskDatasetMetricValueModel
Image-to-Image TranslationCityscapes Labels-to-PhotoFID54.2SPADE + SESAME
Image-to-Image TranslationCityscapes Labels-to-PhotomIoU66SPADE + SESAME
Image-to-Image TranslationADE20K Labels-to-PhotosFID31.9SPADE + SESAME
Image-to-Image TranslationADE20K Labels-to-PhotosmIoU49SPADE + SESAME
Image GenerationCityscapes Labels-to-PhotoFID54.2SPADE + SESAME
Image GenerationCityscapes Labels-to-PhotomIoU66SPADE + SESAME
Image GenerationADE20K Labels-to-PhotosFID31.9SPADE + SESAME
Image GenerationADE20K Labels-to-PhotosmIoU49SPADE + SESAME
1 Image, 2*2 StitchingCityscapes Labels-to-PhotoFID54.2SPADE + SESAME
1 Image, 2*2 StitchingCityscapes Labels-to-PhotomIoU66SPADE + SESAME
1 Image, 2*2 StitchingADE20K Labels-to-PhotosFID31.9SPADE + SESAME
1 Image, 2*2 StitchingADE20K Labels-to-PhotosmIoU49SPADE + SESAME

Related Papers

fastWDM3D: Fast and Accurate 3D Healthy Tissue Inpainting2025-07-17Synthesizing Reality: Leveraging the Generative AI-Powered Platform Midjourney for Construction Worker Detection2025-07-17FashionPose: Text to Pose to Relight Image Generation for Personalized Fashion Visualization2025-07-17A Distributed Generative AI Approach for Heterogeneous Multi-Domain Environments under Data Sharing constraints2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17Beyond Fully Supervised Pixel Annotations: Scribble-Driven Weakly-Supervised Framework for Image Manipulation Localization2025-07-17FADE: Adversarial Concept Erasure in Flow Models2025-07-16CharaConsist: Fine-Grained Consistent Character Generation2025-07-15