TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Methods/LAPGAN

LAPGAN

Computer VisionIntroduced 20005 papers
Source Paper

Description

A LAPGAN, or Laplacian Generative Adversarial Network, is a type of generative adversarial network that has a Laplacian pyramid representation. In the sampling procedure following training, we have a set of generative convnet models {G_0,…,G_KG\_{0}, \dots , G\_{K}G_0,…,G_K}, each of which captures the distribution of coefficients h_kh\_{k}h_k for natural images at a different level of the Laplacian pyramid. Sampling an image is akin to a reconstruction procedure, except that the generative models are used to produce the h_kh\_{k}h_k’s:

I~_k=u(I~_k+1)+h~_k=u(I~_k+1)+G_k(z_k,u(I~_k+1)) \tilde{I}\_{k} = u\left(\tilde{I}\_{k+1}\right) + \tilde{h}\_{k} = u\left(\tilde{I}\_{k+1}\right) + G\_{k}\left(z\_{k}, u\left(\tilde{I}\_{k+1}\right)\right)I~_k=u(I~_k+1)+h~_k=u(I~_k+1)+G_k(z_k,u(I~_k+1))

The recurrence starts by setting I~_K+1=0\tilde{I}\_{K+1} = 0I~_K+1=0 and using the model at the final level G_KG\_{K}G_K to generate a residual image I~_K\tilde{I}\_{K}I~_K using noise vector z_Kz\_{K}z_K: I~_K=G_K(z_K)\tilde{I}\_{K} = G\_{K}\left(z\_{K}\right)I~_K=G_K(z_K). Models at all levels except the final are conditional generative models that take an upsampled version of the current image I~_k+1\tilde{I}\_{k+1}I~_k+1 as a conditioning variable, in addition to the noise vector z_kz\_{k}z_k.

The generative models {G_0,…,G_KG\_{0}, \dots, G\_{K}G_0,…,G_K} are trained using the CGAN approach at each level of the pyramid. Specifically, we construct a Laplacian pyramid from each training image III. At each level we make a stochastic choice (with equal probability) to either (i) construct the coefficients h_kh\_{k}h_k either using the standard Laplacian pyramid coefficient generation procedure or (ii) generate them using $G_{k}:

h~_k=G_k(z_k,u(I_k+1))\tilde{h}\_{k} = G\_{k}\left(z\_{k}, u\left(I\_{k+1}\right)\right)h~_k=G_k(z_k,u(I_k+1))

Here G_kG\_{k}G_k is a convnet which uses a coarse scale version of the image l_k=u(I_k+1)l\_{k} = u\left(I\_{k+1}\right)l_k=u(I_k+1) as an input, as well as noise vector z_kz\_{k}z_k. D_kD\_{k}D_k takes as input h_kh\_{k}h_k or h~_k\tilde{h}\_{k}h~_k, along with the low-pass image l_kl\_{k}l_k (which is explicitly added to h_kh\_{k}h_k or h~_k\tilde{h}\_{k}h~_k before the first convolution layer), and predicts if the image was real or generated. At the final scale of the pyramid, the low frequency residual is sufficiently small that it can be directly modeled with a standard GAN: h~_K=G_K(z_K)\tilde{h}\_{K} = G\_{K}\left(z\_{K}\right)h~_K=G_K(z_K) and D_KD\_{K}D_K only has h_Kh\_{K}h_K or h~_K\tilde{h}\_{K}h~_K as input.

Breaking the generation into successive refinements is the key idea. We give up any “global” notion of fidelity; an attempt is never made to train a network to discriminate between the output of a cascade and a real image and instead the focus is on making each step plausible.

Papers Using This Method

Generative Adversarial Networks Bridging Art and Machine Intelligence2025-02-06Medical Image Generation using Generative Adversarial Networks2020-05-19Generating Highly Realistic Images of Skin Lesions with GANs2018-09-05MelanoGANs: High Resolution Skin Lesion Synthesis with GANs2018-04-12Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks2015-06-18