Generative AI: Statistical and probabilistic principles

Graduate course at UNC on the statistical and probabilistic principles of Generative AI

View the Project on GitHub shankar-bhamidi/gaspp

STOR 89X (Spring 2026): GASPp

Generative AI, Statistical and Probabilistic (p)rinciples

Instructor: Shankar Bhamidi
Meeting time: Tue–Thu, 11:00–12:15
Location: Hanes 107

Note on course logistics: Official course materials, announcements, submissions, and grades will be handled via UNC Canvas.
This public website provides a lightweight “front door” with links to slides/notes/readings (e.g., via Dropbox) and a running schedule.


Course overview

This course explores Generative AI through the lens of probability and statistics, with a particular emphasis on identifying—or questioning—the existence of coherent statistical and probabilistic principles underlying modern generative models.

The course title GASPp reflects this goal:

This is an intentionally exploratory, research-oriented course. Roughly two-thirds of the semester will focus on building foundational understanding of:

The course is not implementation-heavy; the emphasis is on understanding, intuition, reading research papers critically, and identifying research directions.

Official syllabus

More details can be found here

Major acknowledgement

I can only convey my understanding of the material, much of which I learnt from the material in the Resources. In particular I need to mention the wonderful resources developed by Prof. Cosma Shalizi and Prof. Ambuj Tewari. I am largely following in the shadow of these giants.


Learning goals

By the end of the semester, students should:



Announcements


Schedule (living)

Official announcements/submissions are via Canvas. This part of the page is the “living” schedule with links (e.g., Dropbox) to slides, notes, and readings. If the links below do not work, this folder link to all the lectures should hopefully be more stable.

Date Topic Materials
2026-01-08 (Thu) Intro to course. Class expectations. BPE Tokenization. Slides: Lecture 1, P1-49 Reading: notes
2026-01-13 (Tue) Vector semantics and representation learning. Classification and inevitability of logistic regression. Slides: Lecture 1, slides 49 - 114 Reading: notes
2026-01-15 (Thu) Word2vec and node2vec. Toy models. Start of information theory Slides: Lecture 1, slides 114 - end, Lecture 2, slides 1-9 Reading: notes
2026-01-20 (Tue) Basics of information theory Slides: Lecture 2, Slides 9-41, Reading: notes
2026-01-22 (Thu) Data processing inequality, Information theory and MLE and LDP. Entropy rates for stochastic processes Slides: Lecture 2, associated slides,
2026-01-27 (Tue) Classes cancelled owing to snow day.  
2026-01-29 (Thu) Ergodic sequences and entropy. Convexity of KL. Information geometry (maximization of Entropy, ELBO) Slides: Lecture 2, associated slides, Reading: notes
2026-02-03 (Tue) N-gram models and their use Slides: Lecture 3
2026-02-05 (Thu) Finished N-gram models. Started Neural networks. Origin story. Slides: Lecture 3, Lecture 4, Slides 1-20
2026-02-10 (Tue) Neural networks, Kolmogarov and Arnold. Baby implementation in Logistic regression Lecture 4, Slides 21-45
2026-02-12 (Thu) Neural networks and Stochastic gradient descent, Back propogation. Noise contrastive estimation Lectue 4, slides 45-86
2026-02-17 (Tue) Basics of Stochastic gradient descent Lecture 4, slides 95-115
2026-02-19 (Thu) Convergence a.s for stochastic gradient descent, state of the art for high dimensional stochastic gradient descent Lecture 4 till the end
2026-02-24 (Tue) Enter the transformer Lecture 5
2026-02-26 (Thu)    
2026-03-03 (Tue)    
2026-03-05 (Thu)    
2026-03-10 (Tue)    
2026-03-12 (Thu)    

Detailed reading

2026-01-08 reading

Material covered in the lecture

Suggested reading and additional resources

2026-01-13 reading

Material covered in the lecture

Suggested reading and additional resources

2026-01-16 reading

Material covered in the lecture

Suggested reading and additional resources

2026-01-20 reading

Material covered in the lecture

Suggested reading and additional resources

2026-01-29 reading

Material covered in the lecture

Suggested reading and additional resources


Spring 2026 — Tentative plan at the start of the semester (subject to change)

Date Topic Materials
2026-01-08 (Thu) Intro to course. Class expectations. Slides: link · Notes: link
2026-01-13 (Tue) BPE / N-grams / vector semantics / representation learning Slides: link · Reading: link
2026-01-15 (Thu) Information theory: cross-entropy loss and MLE Slides: link · Reading: link
2026-01-20 (Tue) AEP, ergodic theory, Shannon–McMillan–Breiman Slides: link · Reading: link
2026-01-22 (Thu) Markov chains for text; stochastic gradient descent Slides: link · Reading: link
2026-01-27 (Tue) SGD in low and high dimensions Slides: link · Reading: link
2026-01-29 (Thu) SGD continued: neural networks Slides: link · Reading: link
2026-02-03 (Tue) Transformers, attention Slides: link · Reading: link
2026-02-05 (Thu) Transformers, attention (cont.) Slides: link · Reading: link
2026-02-10 (Tue) Transformers and interacting particle systems Slides: link · Reading: link
2026-02-12 (Thu) Prompting as conditioning; RLHF Slides: link · Reading: link
2026-02-17 (Tue) Influence functions; superconcentration; propagation of chaos Slides: link · Reading: link
2026-02-19 (Thu) Interlude: graph rep learning; spectral clustering Slides: link · Reading: link
2026-02-24 (Tue) Interlude: graph neural networks Slides: link · Reading: link
2026-02-26 (Thu) Latent variables I: K-means and EM Slides: link · Reading: link
2026-03-03 (Tue) Latent variables II: VI & ELBO; score matching Slides: link · Reading: link
2026-03-05 (Thu) GANs and autoencoders Slides: link · Reading: link
2026-03-10 (Tue) GANs and autoencoders II Slides: link · Reading: link
2026-03-12 (Thu) Diffusion models I Slides: link · Reading: link