Cover Image for Diffusion Model Meetup & Paper Reading — Diffusion Transformers & Scalable Diffusion Models with Transformers, with Paper and Code
Cover Image for Diffusion Model Meetup & Paper Reading — Diffusion Transformers & Scalable Diffusion Models with Transformers, with Paper and Code
Avatar for AI Scholars
Presented by
AI Scholars
Hosted By
14 Going

Diffusion Model Meetup & Paper Reading — Diffusion Transformers & Scalable Diffusion Models with Transformers, with Paper and Code

Zoom
Registration
Welcome! To join the event, please register below.
About Event

TL;DR
In this session, we’ll walk through one of the landmark papers connecting transformers and diffusion models — “Scalable Diffusion Models with Transformers” (Diffusion Transformer / DiT).​
We’ll unpack why replacing U‑Nets with transformers in diffusion models is a big deal, how the DiT architecture is designed, and what this means for scaling high‑quality image generation.​
This session is part of our ongoing Diffusion Model Paper Reading Group, a friendly, online community across NY, SF, Toronto, and Boston — open to anyone curious about modern generative models.​


👌 Learning requirements

You’ll be fine as long as you’re:

  • Curious about how diffusion models and transformers can be combined

  • Comfortable skimming a technical paper and engaging in discussion

  • Open to learning visually and conceptually (light math, optional code intuition, no coding required during the session)


🗓 Schedule

  • First 60 min: High‑level walkthrough of the “Scalable Diffusion Models with Transformers” paper and core DiT architecture, with Code.

  • Second 30 min: Open group discussion and Q&A to clarify concepts and connect DiTs to earlier diffusion and transformer sessions.​


📚 Pre‑class learning

  • Paper: Scalable Diffusion Models with Transformers (DiT) — arXiv:2212.09748
    https://arxiv.org/pdf/2212.09748

  • Recommended video (concise, < 30 min):
    Scalable Diffusion Models with Transformers — DiT Explanation and Implementation by hu‑po
    https://www.youtube.com/watch?v=aSLDXdc2hkk
    Clear, focused walkthrough of the DiT paper and architecture, ideal if you want a streamlined explanation.

  • Additional video (~1 hour+ deep dive):
    Stanford CS25: Transformers in Diffusion Models for Image Generation and Beyond by Sayak Paul (Hugging Face)
    https://www.youtube.com/watch?v=vXtapCFctTI
    Comprehensive lecture from Stanford’s CS25 series that connects U‑Nets, diffusion models, and Diffusion Transformers in a broader research context.


👥 Speakers

Led by master’s and PhD students in AI, IBM‑style applied AI consultants, and CTO‑level builders from award‑winning AI startups — all experienced in explaining diffusion models and transformer architectures clearly to practitioners.​
The highlight of this session is clarity — by the end, you’ll understand how Diffusion Transformers are structured, why they scale so well, and how they fit into the broader generative model landscape.


🧠 About the Diffusion Model Reading Group

A peer‑led, 5‑month learning journey for engineers, students, researchers, and builders exploring diffusion architectures and modern AI.​

  • No ML background required — just curiosity and some comfort with technical ideas

  • 2–4 hours/week with paper readings, discussions, and final projects

  • Supportive community of people already working in or entering the AI industry

Avatar for AI Scholars
Presented by
AI Scholars
Hosted By
14 Going