Generative Modeling & Probability Foundations
The landscape of generative models, probability distributions, KL divergence, evidence lower bounds, and why learning to reverse noise is the key insight behind diffusion.
A deep technical series demystifying how diffusion models and flow matching actually work — from probability foundations to state-of-the-art generation. Built for engineers who want to understand, not just use.
The landscape of generative models, probability distributions, KL divergence, evidence lower bounds, and why learning to reverse noise is the key insight behind diffusion.
The forward noising process, reverse denoising, the variational bound, noise prediction networks, and the surprisingly simple DDPM training objective.
Score functions as gradients of log-density, score matching, denoising score matching, annealed Langevin dynamics, and noise-conditional score networks.
Continuous-time diffusion as stochastic differential equations, forward and reverse SDEs, the probability flow ODE, and variance-preserving vs variance-exploding formulations.
Neural ODEs, continuous normalizing flows, the flow matching objective, conditional flow matching, rectified flows, and optimal transport paths.
U-Net design for diffusion, Diffusion Transformers (DiT), classifier guidance, classifier-free guidance, text conditioning via cross-attention, and latent diffusion.
DDIM, DPM-Solver, Euler and Heun methods, progressive distillation, consistency models, and how to generate in 1–4 steps instead of 1000.
Image, video, 3D, and audio synthesis, inpainting and editing, discrete diffusion for text, molecular generation, and the convergence of diffusion and flow matching.