Abstract
There is a widely-spread claim that GANs are difficult to train, and GAN architectures in the literature are littered with empirical tricks. We provide evidence against this claim and build a modern GAN baseline in a more principled manner. First, we derive a well-behaved regularized relativistic GAN loss that addresses issues of mode dropping and non-convergence that were previously tackled via a bag of ad-hoc tricks. We analyze our loss mathematically and prove that it admits local convergence guarantees, unlike most existing relativistic losses. Second, our new loss allows us to discard all ad-hoc tricks and replace outdated backbones used in common GANs with modern architectures. Using StyleGAN2 as an example, we present a roadmap of simplification and modernization that results in a new minimalist baseline -- R3GAN. Despite being simple, our approach surpasses StyleGAN2 on FFHQ, ImageNet, CIFAR, and Stacked MNIST datasets, and compares favorably against state-of-the-art GANs and diffusion models.
Community
Can GANs beat diffusion in 2025? Yes!
This paper describes how to train GAN models more stably than prior work. The authors are able to modernize the ConvNet architecture of GANs while maintaining the performance and equilibrium of the model. This newer, more stable training strategy allows the GAN to train for longer, matching a similar number of training steps as a diffusion models. Once GANs are trained long enough, with a powerful enough architecture, they can outperform diffusion models, and serve as better, faster, smaller models.
We created a deep dive video for this paper: https://www.youtube.com/watch?v=tMJx2s8YsZs. Happy learning🤓 and stretching💪 together!✨🐸
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Mechanisms of Generative Image-to-Image Translation Networks (2024)
- A Framework for Image Synthesis Using Supervised Contrastive Learning (2024)
- Diverse Rare Sample Generation with Pretrained GANs (2024)
- Enhancing Few-Shot Learning with Integrated Data and GAN Model Approaches (2024)
- Jet: A Modern Transformer-Based Normalizing Flow (2024)
- Parallelly Tempered Generative Adversarial Networks (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
First author is incorrectly linked at the moment: https://huggingface.co./NickBVC is the proper HF account. Working on rectifying it.
Models citing this paper 5
Browse 5 models citing this paperDatasets citing this paper 0
No dataset linking this paper