Curated Track
Math to ML
You know calculus, linear algebra, and some probability. This path takes you from mathematical foundations through learning theory to understanding why ML models work, when they fail, and what the theorems actually say.
This is a curated default sequence. Use it as a strong spine, then skip ahead when the prerequisites are already solid.
Phase 1: Core Foundations
Make sure your mathematical foundations are solid. If you can do these, you are ready for learning theory.
Phase 2: Concentration and Estimation
The tools that make learning theory work. Every generalization bound uses these.
Phase 3: Learning Theory
Now you can understand why ML works. This is the core of the theoretical spine.
Phase 4: Optimization and Methods
How models are actually trained. The bridge between theory and practice.
Phase 5: Modern Deep Learning Theory
Why overparameterized networks generalize, and what classical theory gets wrong.
Not sure where to start?
If Phase 1 feels too basic, skip to Phase 2 or 3. If Phase 3 feels too hard, go back to Phase 2. The prerequisite sidebar on each page shows what you need.
Or search for a specific concept →