Shengyu Zhao, Jonathan Cui, Yilun Sheng, Yue Dong, Xiao Liang, Eric I Chang, Yan Xu
Numerous task-specific variants of conditional generative adversarial networks have been developed for image completion. Yet, a serious limitation remains that all existing algorithms tend to fail when handling large-scale missing regions. To overcome this challenge, we propose a generic new approach that bridges the gap between image-conditional and recent modulated unconditional generative architectures via co-modulation of both conditional and stochastic style representations. Also, due to the lack of good quantitative metrics for image completion, we propose the new Paired/Unpaired Inception Discriminative Score (P-IDS/U-IDS), which robustly measures the perceptual fidelity of inpainted images compared to real images via linear separability in a feature space. Experiments demonstrate superior performance in terms of both quality and diversity over state-of-the-art methods in free-form image completion and easy generalization to image-to-image translation. Code is available at https://github.com/zsyzzsoft/co-mod-gan.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Image Generation | FFHQ 512 x 512 | FID | 3.7 | CoModGAN |
| Image Generation | Places2 | FID | 2.92 | CoModGAN |
| Image Generation | Places2 | P-IDS | 19.64 | CoModGAN |
| Image Generation | Places2 | U-IDS | 35.78 | CoModGAN |
| Image Generation | CelebA-HQ | FID | 5.65 | CoModGAN |
| Image Generation | CelebA-HQ | P-IDS | 11.23 | CoModGAN |
| Image Generation | CelebA-HQ | U-IDS | 22.54 | CoModGAN |
| Image Inpainting | FFHQ 512 x 512 | FID | 3.7 | CoModGAN |
| Image Inpainting | Places2 | FID | 2.92 | CoModGAN |
| Image Inpainting | Places2 | P-IDS | 19.64 | CoModGAN |
| Image Inpainting | Places2 | U-IDS | 35.78 | CoModGAN |
| Image Inpainting | CelebA-HQ | FID | 5.65 | CoModGAN |
| Image Inpainting | CelebA-HQ | P-IDS | 11.23 | CoModGAN |
| Image Inpainting | CelebA-HQ | U-IDS | 22.54 | CoModGAN |