Improved Training of Wasserstein GANs

Ishaan Gulrajani, Faruk Ahmed, Martin Arjovsky, Vincent Dumoulin, Aaron Courville

Add an identifier
Abstract:Generative Adversarial Networks (GANs) are powerful generative models, but suffer from training instability. The recently proposed Wasserstein GAN (WGAN) makes progress toward stable training of GANs, but sometimes can still generate only low-quality samples or fail to converge. We find that these problems are often due to the use of weight clipping in WGAN to enforce a Lipschitz constraint on the critic, which can lead to undesired behavior. We propose an alternative to clipping weights: penalize the n ...
What's meant by "Residual block"?
0
2
Entering edit mode
2.8 years ago
Andy 46

In Appendix F, the authors say "The generator and critic are residual networks; we use pre-activation residual blocks with two 3 × 3 convolutional layers each and ReLU nonlinearity."

screenshot of architecture description

Based on this, I'm thinking something like

def resblock(x, training=False):
    y = conv1(x)
    y = bn1(x, training=training)
    y = relu(x)

    y = conv2(x)
    y = bn2(x, training=training)

    return relu(y + matmul(W, x))

where W is some trainable weight matrix sized correctly to account for whatever padding is being used in the conv layers. I'm not really sure what's meant by "pre-activation residual blocks" though and generally, I'm not sure if I've got the implementation quite right.

This is also related to the paper arXiv:1802.05957

gan • 851 views

Login before adding your answer.

Traffic: 1 users visited in the last hour