Modern Transformer Architecture: A Curated YouTube Course(x.com)
Jia-Bin Huang (UMD) curated a short YouTube course on Transformer architecture: attention, positional encodings, vision transformers, and recent variants. If you work in CV and feel like you picked up transformers by osmosis rather than really understanding them, this is a good way to fill the gaps. Links in the thread, in order.
Flow Matching and Diffusion Models from Scratch: Free Lecture Notes(arxiv.org)
Lecture notes from Peter Holderrieth and Ezra Erives covering flow matching and diffusion models from scratch. No background in generative modeling assumed. Goes from the math up to current state of the art. Reading the original papers in sequence (DDPM, DDIM, flow matching) works, but each one assumes you got something from the previous. These notes don't make that assumption.
LLM Architecture Gallery: Every Major Architecture in One Place(sebastianraschka.com)
Sebastian Raschka collected architecture diagrams for most of the major LLM families in one place: GPT, BERT, T5, LLaMA variants, Mistral, Gemma. When a paper says it builds on LLaMA-2 with GQA and you want to know what that actually looks like, this is faster than digging through GitHub readmes.
A Complete Guide to Learning ML from YouTube (13 Courses)(x.com)
Sanju Sinha put this together from what he actually watched during his PhD, in the order he'd recommend watching them. Not just a list of popular channels, but sequenced to build on each other. Good to share with anyone just getting started in the lab.