Please report any queries concerning the funding data grouped in the sections named "Externally Awarded" or "Internally Disbursed" (shown on the profile page) to
your Research Finance Administrator. Your can find your Research Finance Administrator at https://www.ucl.ac.uk/finance/research/rs-contacts.php by entering your department
Please report any queries concerning the student data shown on the profile page to:
Email: portico-services@ucl.ac.uk
Help Desk: http://www.ucl.ac.uk/ras/portico/helpdesk
Email: portico-services@ucl.ac.uk
Help Desk: http://www.ucl.ac.uk/ras/portico/helpdesk
Publication Detail
Demystifying MMD GANs
-
Publication Type:Conference
-
Authors:BiĆkowski M, Sutherland DJ, Arbel M, Gretton A
-
Publication date:03/05/2018
-
Pagination:1, 11
-
Name of conference:International Conference on Learning Representations
-
Conference place:Vancouver, Canada
-
Conference start date:30/04/2018
-
Conference finish date:03/05/2018
-
Keywords:stat.ML, stat.ML, cs.LG
-
Author URL:
-
Publisher URL:
-
Notes:Published at ICLR 2018: https://openreview.net/forum?id=r1lUOzWCW . v3: final conference version: ResNet CelebA results, more on FID/KID comparisons
Abstract
We investigate the training and performance of generative adversarial
networks using the Maximum Mean Discrepancy (MMD) as critic, termed MMD GANs.
As our main theoretical contribution, we clarify the situation with bias in GAN
loss functions raised by recent work: we show that gradient estimators used in
the optimization process for both MMD GANs and Wasserstein GANs are unbiased,
but learning a discriminator based on samples leads to biased gradients for the
generator parameters. We also discuss the issue of kernel choice for the MMD
critic, and characterize the kernel corresponding to the energy distance used
for the Cramer GAN critic. Being an integral probability metric, the MMD
benefits from training strategies recently developed for Wasserstein GANs. In
experiments, the MMD GAN is able to employ a smaller critic network than the
Wasserstein GAN, resulting in a simpler and faster-training algorithm with
matching performance. We also propose an improved measure of GAN convergence,
the Kernel Inception Distance, and show how to use it to dynamically adapt
learning rates during GAN training.
› More search options
UCL Researchers