Google Research

Consistency Regularization for Generative Adversarial Networks

International Conference on Learning Representations (2020)


Generative Adversarial Networks are plagued by training instability, despite considerable research effort. Progress has been made on this topic, but many of the proposed interventions are complicated, computationally expensive, or both. In this work, we propose a simple and effective training stabilizer based on the notion of Consistency Regularization - a popular technique in the Semi-Supervised Learning literature. In particular, we augment data passing into the GAN discriminator and penalize the sensitivity of the penultimate layer of the discriminator to these augmentations. This regularization increases the robustness of the discriminator to input perturbations and demonstrably reduces memorization of the training data. We conduct a series of ablation studies to demonstrate that consistency regularization is compatible with various GAN architectures and loss functions. Finally, we show that applying consistency regularization to GANs improves state-of-the-art FID scores on the ImageNet-2012 data set. Our code is open-sourced at \textbf{URL blinded for peer review}.

Learn more about how we do research

We maintain a portfolio of research projects, providing individuals and teams the freedom to emphasize specific types of work