What is: Introspective Adversarial Network?
Source | Neural Photo Editing with Introspective Adversarial Networks |
Year | 2000 |
Data Source | CC BY-SA - https://paperswithcode.com |
The Introspective Adversarial Network (IAN) is a hybridization of GANs and VAEs that leverages the power of the adversarial objective while maintaining the VAE’s efficient inference mechanism. It uses the discriminator of the GAN, , as a feature extractor for an inference subnetwork, , which is implemented as a fully-connected layer on top of the final convolutional layer of the discriminator. We infer latent values for reconstruction and sample random values from a standard normal for random image generation using the generator network, .
Three distinct loss functions are used:
- , the L1 pixel-wise reconstruction loss, which is preferred to the L2 reconstruction loss for its higher average gradient.
- , the feature-wise reconstruction loss, evaluated as the L2 difference between the original and reconstruction in the space of the hidden layers of the discriminator.
- , the ternary adversarial loss, a modification of the adversarial loss that forces the discriminator to label a sample as real, generated, or reconstructed (as opposed to a binary real vs. generated label).
Including the VAE’s KL divergence between the inferred latents and the prior , the loss function for the generator and encoder network is thus:
Where the terms weight the relative importance of each loss. We set to 3 and leave the other terms at 1. The discriminator is updated solely using the ternary adversarial loss. During each training step, the generator produces reconstructions (using the standard VAE reparameterization trick) from data and random samples , while the discriminator observes as well as the reconstructions and random samples, and both networks are simultaneously updated.