06-reference

dami defi 12 graduate ml textbooks curation

Wed Apr 29 2026 20:00:00 GMT-0400 (Eastern Daylight Time) ·reference ·source: X (Dami-Defi article, shared by founder via iMessage) ·by Dami-Defi (@DamiDefi on X)

12 free graduate ML textbooks — curation source for the RDCO bookshelf

Why this is in the vault

Founder shared the article on 2026-04-30 with the framing “I think our setup is more advanced” — meaning the article’s “upload PDFs into a Claude Project as ambient context” workflow is naive vs. the bookshelf architecture proposed in 2026-04-30-bookshelf-source-material-architecture-gap. But the curation list itself is genuinely valuable as a starter set for the ML-fundamentals domain when bookshelf scaffold lands. Filing for that.

The 12 textbooks (with corrected attribution)

The article’s “MIT graduate-level AI textbooks” framing is loose marketing — only ~2-3 of the 12 are actually MIT. The real story: these are the canonical free graduate ML textbooks across the field, regardless of host institution.

Foundations

  1. Foundations of Machine Learning — Mohri, Rostamizadeh, Talwalkar
  2. Understanding Deep Learning — Simon J.D. Prince
    • Institution: University of Bath
    • URL: https://udlbook.github.io/udlbook
    • What it covers: clearest in-print explanation of how neural nets process and weight information
  3. Machine Learning Systems — Vijay Janapa Reddi (lead) + community
    • Institution: Harvard / MIT (textbook is part of the broader MLSysBook collaboration; “MIT curriculum” framing in the article is closest to accurate here)
    • URL: https://mlsysbook.ai
    • What it covers: bridges theory to production; how models are built, deployed, and fail in real environments

Advanced techniques

  1. Algorithms for Decision Making — Kochenderfer, Wheeler, Wray
  2. Deep Learning — Goodfellow, Bengio, Courville
    • Publisher: MIT Press (this is the only legitimate “MIT” thread in the article’s framing)
    • URL: https://www.deeplearningbook.org
    • What it covers: the original deep learning textbook; dense but foundational

Reinforcement learning

  1. Reinforcement Learning: An Introduction — Sutton & Barto
  2. Distributional Reinforcement Learning — Bellemare, Dabney, Rowland
    • Institution: DeepMind / Mila / Google Brain
    • URL: https://www.distributional-rl.org
    • What it covers: shifts the frame from expected rewards to full probability distributions over outcomes
  3. Multi-Agent Reinforcement Learning — Albrecht, Christianos, Schäfer
    • Institution: University of Edinburgh
    • URL: https://marl-book.com
    • What it covers: how multiple decision-making agents learn to interact in shared environments
  4. Decision Making Under Uncertainty (long-game) — Kochenderfer

Ethics + probability

  1. Fairness and Machine Learning — Barocas, Hardt, Narayanan
    • Institutions: Microsoft Research / UC Berkeley / Princeton
    • URL: https://fairmlbook.org
    • What it covers: where models break down and where bias enters; required reading for knowing when NOT to trust model output
  2. Probabilistic Machine Learning: An Introduction — Kevin Murphy
  3. Probabilistic Machine Learning: Advanced Topics — Kevin Murphy
    • Institution: Google Research
    • URL: https://probml.github.io/book2.html
    • What it covers: causal inference, decision making under uncertainty, generative models at graduate level

Mapping against Ray Data Co

As a bookshelf starter set

When the bookshelf scaffold (2026-04-30-bookshelf-source-material-architecture-gap) lands, this list is a clean candidate for the ML-fundamentals domain. Recommended starter triad (per the article’s own claim, which is defensible):

Adding the full 12 takes ~5-10GB of disk and one afternoon of ingestion work. Defer until founder greenlights the bookshelf scaffold.

Why our (proposed) system is structurally more advanced than the article’s

DimensionDami-Defi patternRDCO bookshelf proposal
PersistencePer-Project context windowFilesystem + QMD collection, indefinite
Retrieval granularityWhole-book ambient contextPassage-level via QMD vector + lexical
Citation chainNone (Claude paraphrases from context)Required: passage + page/timestamp + URL
Synthesis vs source separationMixedTwo distinct QMD collections (vault vs source-material)
Cross-Project reuseNone (Project-bounded)Universal across all skills + sessions
Curation disciplineImplicit (whatever fits in upload)Explicit (per-domain canonical list, founder-judged)
Cost to maintainZero ongoingLow (incremental ingestion via /save-to-bookshelf skill)

Both work. Dami’s is right for “I want to test this in an afternoon.” Ours is right for “I’m running an agent-native company and need this as durable infrastructure.”

Why filing this curation matters even before bookshelf-build

Even if the bookshelf scaffold takes another month to land, this curation note is useful as:

Caveats on the article’s framing

The article (Dami-Defi, X) is heavily marketing-flavored:

What’s actually load-bearing in the article: