10-part technical series

Deep Learning
from the Inside Out

Every architecture broken down to its weight matrices, computation diagrams, and FLOPs. No hand-waving -- just the math, the shapes, and the intuition.

Hiva Mohammadzadeh

Hiva Mohammadzadeh

Stanford MSCS UC Berkeley EECS

MS in Computer Science at Stanford. Previously UC Berkeley EECS. Researching efficient NLP, LLM architectures, and model compression.

📚
These blog posts are based on presentations and tutorials I created in 2023 to survey Large Language Models, benchmarks, normalization techniques, and activation functions, along with my CS199 Supervised Independent Study at UC Berkeley EECS where I surveyed 11 foundational papers from "Attention is All You Need" through GPT-4. Now expanded into this blog series.
10 Deep Dives
4 LLM Architectures

Foundations

Model Architectures

Techniques