The GAN Era (2014–2021)
GANs dominated image generation from 2014 to ~2021, producing increasingly photorealistic results. They proved that neural networks could create, not just classify. But their training instability, mode collapse, and difficulty with text conditioning led to their gradual replacement by diffusion models for most generation tasks. GANs remain relevant for real-time applications (super-resolution, video enhancement) where diffusion’s slow sampling is prohibitive.
The connection: GANs taught us that adversarial training produces sharp outputs, and that latent spaces can encode meaningful structure. These ideas influenced diffusion models, contrastive learning, and even LLM alignment (RLHF uses a reward model reminiscent of a discriminator). Next: Regularization & Practical Training.
GAN Timeline
// GAN milestones
2014: Original GAN (Goodfellow)
2016: DCGAN (convolutional GANs)
2017: WGAN (Wasserstein distance)
2017: Pix2Pix, CycleGAN
2018: Progressive GAN (1024×1024)
2019: StyleGAN (photorealistic faces)
2019: BigGAN (class-conditional, ImageNet)
2020: StyleGAN2 (artifact-free)
2021: Diffusion models begin to dominate
2022: DALL-E 2, Stable Diffusion replace GANs