layout | title | permalink | ordinal |
---|---|---|---|
page |
Unrolled GANs |
/unrolled/ |
4 |
Unrolled GANs are a variation of GANs. The problem they try to resolve is -
They try to stabilize the learning of GANs and thus, leads to convergence. Also, they try to eliminate the problem of mode collapse in vanilla GANs wherein there is little to no variation in the images generated by the generator.
Consider the loss function
Let optimal parameters be $\theta_G^* $ and $\theta_D^* $ for generator and discriminator respectively.
Then, using the loss equation of vanilla GANs -
The main idea presented in this paper is that, the generator just tries to fool the discriminator at this point of time. However, this could lead to oscillation.
An example of such oscillation -
Consider two generator states
Now, after training the discriminator, let the discriminator go to state
Let's say that generator at state
This could lead to an endless oscillation.
So, to avoid such as oscillation, the idea is to train the generator with respect to the approximate discriminator, a few steps into the future.
Consider the following recursive definintion -
i.e. the next value of
So, this is an approximation of the future discriminator.
For training the generator, we use the
where
This means that the generator is updated using the approximate discriminator function from the future.
However, the discriminator is updated in the same way as vanilla GANs, i.e. using only the current generator. So, the update equations are -
Figure 1 represents the procedure for unrolling in 2 steps.
However, for updating the discriminator, only gradients from
A model was trained on a 2D mixture of 8 Gaussians arranged in a circle.
In case of Vanilla GANs, there is a continuous oscillation, which is similar to the example of oscillation given above (only for a larger cycle). Also, there is a mode collapse, since at no time does the generator know about all possibile transformations of latent vector to images. It only knows the transformation corresponding to 1 Gaussian.
However, in case of unrolled GANs, it finally converges and the generator has a complete picture of what the possible transformations are, thus eliminating the problem of mode collapse.