Gaurav Parmar, Dacheng Li, Kwonjoon Lee, Zhuowen Tu
We present a new generative autoencoder model with dual contradistinctive losses to improve generative autoencoder that performs simultaneous inference (reconstruction) and synthesis (sampling). Our model, named dual contradistinctive generative autoencoder (DC-VAE), integrates an instance-level discriminative loss (maintaining the instance-level fidelity for the reconstruction/synthesis) with a set-level adversarial loss (encouraging the set-level fidelity for there construction/synthesis), both being contradistinctive. Extensive experimental results by DC-VAE across different resolutions including 32x32, 64x64, 128x128, and 512x512 are reported. The two contradistinctive losses in VAE work harmoniously in DC-VAE leading to a significant qualitative and quantitative performance enhancement over the baseline VAEs without architectural changes. State-of-the-art or competitive results among generative autoencoders for image reconstruction, image synthesis, image interpolation, and representation learning are observed. DC-VAE is a general-purpose VAE model, applicable to a wide variety of downstream tasks in computer vision and machine learning.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Image Generation | STL-10 | FID | 41.9 | DC-VAE |
| Image Generation | STL-10 | Inception score | 8.1 | DC-VAE |
| Image Generation | LSUN Bedroom 128 x 128 | FID | 14.3 | DC-VAE |
| Image Generation | CelebA 128x128 | FID | 19.9 | DC-VAE |
| Image Generation | CelebA-HQ 256x256 | FID | 15.81 | DC-VAE |
| Image Generation | CIFAR-10 | FID | 17.9 | DC-VAE |