Schedule Overview

This is a 7-week half-semester course (Mini 3) meeting Tuesdays and Thursdays for 80 minutes each. The schedule below is tentative and subject to changes.


Office Hours

Kelly hosts regular office hours every week in-person in Gates and virtually on Discord.

  • In-person: Wednesdays 1:00 PM - 2:00 PM, Gates 8th Floor common area near the printer
  • Virtual: Fridays 11:00 AM - 12:00 PM, Discord

Krish also hosts regular office hours in-person in Gates.

  • In-person: Tuesdays 4:00 PM - 5:00 PM, Gates 8th Floor common area near the printer

Lecture Schedule

Below is the tentative schedule of the course (subject to changes).

Lecture Date Topic Resources Deliverables
1 01/13 Basics of Probabilistic & Generative Modeling
📖 View readings
2 01/15 Denoising Diffusion Models
📖 View readings
3 01/16 Guest Lecture (Modal): How to train & serve your models on Modal
4 01/20 Score-Based Models
📖 View readings
  • Quiz 1
5 01/22 Flow Matching
📖 View readings
  • HW 1 (15%) Due 01/24 Sat
  • Quiz 2
6 01/27 The Design Space of Diffusion Models & Solvers for Fast Sampling
📖 View readings
7 01/29 Guidance & Controllable Generation
📖 View readings
  • Quiz 3
8 02/03 Guest Lecture: Q&A with Max Simchowitz, Diffusion & Flow for Robotics, Control & Decision Making
9 02/05 SOTA Diffusion/Flow Models for Text-to-Image Generation
📖 View readings
  • Quiz 4
  • HW 2 (15%) Due 02/05 Thur
10 02/10 Distillation, Consistency Models & Flow Maps
📖 View readings
  • Quiz 5
11 02/12 Guest Lecture: Linqi (Alex) Zhou from Luma AI
  • HW 3 (20%) Due 02/15 Sun
12 02/17 Discrete Diffusion & Masked Diffusion
📖 View readings
  • Quiz 6
13 02/19 Discrete Flow Matching & Edit Flow
📖 View readings
  • Quiz 7
14 02/24 No Class
  • Final Presentation (15%) Poster submission due 02/25 Wed
15 02/26 Final Poster Presentation
  • HW 4 (20%) Due 02/27 Fri

Readings and Resources by Lecture

Below are the related papers and tutorials for each lecture. All readings are optional and meant to be additional resources for you to deepen your understanding. The reading list will be updated throughout the class.

Lecture 1: Basics of Probabilistic & Generative Modeling

Tutorials

Papers

Lecture 2: Denoising Diffusion Models

Tutorials

Papers

Lecture 4: Score-Based SDEs

Tutorials

Papers

Lecture 5: Flow Matching

Tutorials

Papers

Lecture 6: The Design Space of Diffusion Models & Solvers for Fast Sampling

Papers

Lecture 7: Guidance & Controllable Generation

Papers

Lecture 9: SOTA Diffusion/Flow Models for Text-to-Image Generation

Papers

Lecture 10: Distillation, Consistency Models & Flow Maps

Papers

Lecture 12: Discrete Diffusion & Masked Diffusion

Tutorials

Papers

  • Structured Denoising Diffusion Models in Discrete State-Spaces
    Jacob Austin, Daniel D. Johnson, Jonathan Ho, Daniel Tarlow, Rieck van den Berg
    The D3PM paper; introduces a family of discrete diffusion models with structured transition matrices including absorbing (mask), uniform, and embedding-based diffusion
  • A Continuous Time Framework for Discrete Denoising Models
    Andrew Campbell, Joe Benton, Valentin De Bortoli, Thomas Rainforth, George Deligiannidis, Arnaud Doucet
    Extends discrete diffusion to continuous time using Continuous Time Markov Chains (CTMCs), enabling more principled training and sampling
  • Discrete Diffusion Modeling by Estimating the Ratios of the Data Distribution
    Aaron Lou, Chenlin Meng, Stefano Ermon
    The SEDD paper; introduces score entropy as a training objective for discrete diffusion by estimating ratios of the data distribution, analogous to score matching in continuous diffusion
  • Simple and Effective Masked Diffusion Language Models
    Subham Sekhar Sahoo, Marianne Arriola, Yair Schiff, Aaron Gokaslan, Edgar Marroquin, Justin T Chiu, Alexander Rush, Volodymyr Kuleshov
    Proposes MDLM, a simple masked diffusion language model with an efficient training objective and absorbing-state noise schedule that matches or outperforms autoregressive models on language benchmarks
  • Simplified and Generalized Masked Diffusion for Discrete Data
    Jiaxin Shi, Kehang Han, Zhe Wang, Arnaud Doucet, Michalis K. Titsias
    Unifies and simplifies masked diffusion models, showing that a simple masked diffusion objective generalizes prior work and yields strong performance on text generation
  • LLaDA: Large Language Diffusion with mAsking
    Shen Nie, Fengqi Zhu, Zebin You, Xiaolu Zhang, Jingyang Ou, Jun Hu, Jun Zhou, Yankai Lin, Ji-Rong Wen, Chongxuan Li
    A masked diffusion language model trained from scratch at scale that matches LLaMA3 8B in instruction following, demonstrating the viability of discrete diffusion for LLMs

Lecture 13: Discrete Flow Matching & Edit Flow

Papers