New Self-paced AI courses — learn ML, deep learning, and agents on your schedule. Enroll free

Generative modeling

Generative Adversarial Nets

Goodfellow et al. · NeurIPS 2014

Paper PDF

Open in new tab

Fetching research paper

Downloading PDF from the archive

If the viewer is blank (blocked by the publisher or your network), use Open in new tab. Scrolling inside the frame moves through the PDF pages when embedding is supported.

Reading map

These notes are written in plain language for this specific paper—so you can grasp the ideas before you wrestle with the authors’ formal wording. Use the button to open the PDF near the matching section (approximate page; Chromium-style viewers support #page=, otherwise we open a new tab).

Problem statement & goal

Learning the full distribution of images (or other data) with classic models was hard. The authors asked: can we learn to generate realistic samples without writing down an explicit density—by instead training two networks in a game?

Methodology & architecture

There are two networks: a generator that fakes data and a discriminator that guesses real vs. fake. Training is a minimax game: improve G to fool D, and D to spot fakes. No closed-form likelihood—just gradients through this two-player setup.

Datasets & benchmarks

Experiments start on MNIST, TFD, and CIFAR-10—small, standard sets where you can see samples and compare visually to baselines. The point is to show the idea works before scaling up.

Results & evaluation metrics

Success is partly qualitative (do samples look real?) and partly quantitative where metrics exist. Students should notice instability: training can oscillate; the paper is honest that this is a research prototype, not a plug-and-play recipe.

Limitations & future work

Mode collapse (generator outputs a few repeated modes) and training balance between G and D are classic GAN pain points. The paper opens the door; later work (WGAN, improved objectives) addresses what’s fragile here.

Reproducibility

You get network architectures, objectives, and training procedure, but not a modern code drop in the original paper. Enough to reimplement the idea in a course project; production GAN training usually needs tricks from later papers.

What to focus on

Eight highlights per paper—why each part matters before you read dense notation and proofs.

Why GANs existed

Modeling high-dimensional data densities explicitly is hard. GANs sidestep writing p(x) in closed form by pitting two networks: one fakes data, one judges real vs. fake.

Minimax game

Training is not ordinary loss minimization—it's a two-player objective. Understanding who moves first in practice (alternate updates) matters for stability intuition.

Generator vs. discriminator

If D is too weak, G cheats easily; if D is too strong, gradients vanish. The paper exposes how delicate that balance is—still a core GAN teaching point.

No likelihood, still samples

You don't evaluate an exact log-likelihood during training. Samples are the product—so qualitative inspection and downstream metrics became central to generative evaluation.

Mode collapse

G may output a few repeated modes that fool D. Spotting collapse in figures connects to why later papers changed objectives (e.g., Wasserstein-style losses).

Small-dataset demos

MNIST, faces (TFD), and CIFAR show feasibility before big-GAN era. Compare visual quality to today's diffusion outputs to calibrate progress.

Position vs. VAEs & RBMs

Related-work framing explains trade-offs: approximate inference in VAEs vs. adversarial training here. Useful for interviews and literature surveys.

Lineage to today

GANs influenced image synthesis for years; diffusion now dominates many domains—but adversarial ideas still appear in discriminators, critics, and hybrid systems.

← Back to Research Lab