TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Aggregated Contextual Transformations for High-Resolution ...

Aggregated Contextual Transformations for High-Resolution Image Inpainting

Yanhong Zeng, Jianlong Fu, Hongyang Chao, Baining Guo

2021-04-03Vocal Bursts Intensity PredictionImage InpaintingTexture Synthesis
PaperPDFCodeCode(official)

Abstract

State-of-the-art image inpainting approaches can suffer from generating distorted structures and blurry textures in high-resolution images (e.g., 512x512). The challenges mainly drive from (1) image content reasoning from distant contexts, and (2) fine-grained texture synthesis for a large missing region. To overcome these two challenges, we propose an enhanced GAN-based model, named Aggregated COntextual-Transformation GAN (AOT-GAN), for high-resolution image inpainting. Specifically, to enhance context reasoning, we construct the generator of AOT-GAN by stacking multiple layers of a proposed AOT block. The AOT blocks aggregate contextual transformations from various receptive fields, allowing to capture both informative distant image contexts and rich patterns of interest for context reasoning. For improving texture synthesis, we enhance the discriminator of AOT-GAN by training it with a tailored mask-prediction task. Such a training objective forces the discriminator to distinguish the detailed appearances of real and synthesized patches, and in turn, facilitates the generator to synthesize clear textures. Extensive comparisons on Places2, the most challenging benchmark with 1.8 million high-resolution images of 365 complex scenes, show that our model outperforms the state-of-the-art by a significant margin in terms of FID with 38.60% relative improvement. A user study including more than 30 subjects further validates the superiority of AOT-GAN. We further evaluate the proposed AOT-GAN in practical applications, e.g., logo removal, face editing, and object removal. Results show that our model achieves promising completions in the real world. We release code and models in https://github.com/researchmm/AOT-GAN-for-Inpainting.

Results

TaskDatasetMetricValueModel
Image GenerationPlaces2FID10.64AOT GAN
Image GenerationPlaces2P-IDS3.07AOT GAN
Image GenerationPlaces2U-IDS19.92AOT GAN
Image InpaintingPlaces2FID10.64AOT GAN
Image InpaintingPlaces2P-IDS3.07AOT GAN
Image InpaintingPlaces2U-IDS19.92AOT GAN

Related Papers

RePaintGS: Reference-Guided Gaussian Splatting for Realistic and View-Consistent 3D Scene Inpainting2025-07-11MTADiffusion: Mask Text Alignment Diffusion Model for Object Inpainting2025-06-30Consistent Zero-shot 3D Texture Synthesis Using Geometry-aware Diffusion and Temporal Video Models2025-06-263DeepRep: 3D Deep Low-rank Tensor Representation for Hyperspectral Image Inpainting2025-06-20Hunyuan3D 2.5: Towards High-Fidelity 3D Assets Generation with Ultimate Details2025-06-19TexTailor: Customized Text-aligned Texturing via Effective Resampling2025-06-12EmbodiedGen: Towards a Generative 3D World Engine for Embodied Intelligence2025-06-12Geological Field Restoration through the Lens of Image Inpainting2025-06-05