06-reference

dwarkesh karpathy ghosts not animals

Thu Oct 16 2025 20:00:00 GMT-0400 (Eastern Daylight Time) ·reference ·source: Dwarkesh Patel (YouTube) ·by Dwarkesh Patel + Andrej Karpathy
ai-scalingagentscontinual-learningagi-timelinesrl-vs-imitationkarpathydwarkeshcognitive-corenanochat

Andrej Karpathy — “We’re summoning ghosts, not building animals” (Dwarkesh Patel)

Why this is in the vault

Karpathy is the most credentialed sober voice currently arguing publicly that the “year of agents” framing is wrong and that we are roughly a decade — not 1-2 years — from the kind of AI that can reliably do an intern’s work. He’s not a permabear (he started OpenAI, ran Tesla Autopilot, just wrote nanochat) and he’s not selling a competing product. When he says GPT-5 Pro is “slop” inside his own repo and that the labs are “trying to pretend like this is amazing, and it’s not” — that lines up exactly with Sanity Check’s editorial spine: trust what builders say when they’re not in fundraising mode. This is also the canonical articulation of the “ghosts vs animals” frame that makes the imitation-learning critique tractable.

The core argument

Decade of agents, not year of agents. Karpathy explicitly walks back the agentic-AI hype as “over-prediction.” The three things missing — multimodality, computer use, and continual learning — are tractable but each requires years of integration work. His 15 years of seeing AI predictions is the basis for the decade estimate; he doesn’t claim more.

We’re building ghosts, not animals. This is the frame the title comes from. Animals are the product of evolution: huge amounts of behavior are pre-baked into the weights via DNA before any learning happens (a zebra runs minutes after birth — that’s not RL). LLMs are the product of imitation of internet text — they’re “ethereal spirit entities” mimicking humans. Different optimization process, different starting point. Pre-training is “crappy evolution” — the practical version we can actually run. He pushes back hard on Sutton’s “build animals” frame: not because it’s wrong in principle, but because we don’t have the substrate for it.

Cognitive core hypothesis. Pre-training does two things simultaneously: it gives the model knowledge AND it gives the model intelligence (the algorithms for in-context learning, etc.). The knowledge is partly holding the intelligence back — agents are bad at “going off the data manifold.” The research direction he’s most interested in: strip the knowledge, keep the cognitive core. A small intelligent entity that knows it doesn’t know X and can go look it up.

Coding agents as honest signal. Most of the Karpathy section on coding is the most quotable in the entire interview. He’s an autocomplete-tab loyalist, not a vibe-coder. On nanochat (intentionally unique repo, not boilerplate): the models kept trying to use DDP when he’d rolled his own gradient sync, kept inserting try/catch defensive bloat, kept using deprecated APIs, kept misunderstanding intent. “They’re way too over-defensive. They keep trying to make a production code base, and I have a bunch of assumptions in my code, and it’s okay.” GPT-5 Pro as the “oracle” he occasionally consults with full repo context — surprisingly good vs a year ago, but still not net useful for novel work.

The recursive-self-improvement objection. Dwarkesh frames the AI-2027-style argument: if you have Claude Code making CRUD apps, imagine a million of these inside OpenAI tweaking architectures. Karpathy: that’s exactly the asymmetry — the models are bad at code that has never been written before, which is what AI research IS. This is his main mechanical reason for longer timelines.

Compiler analogy. Programming has had decades of productivity wins (compilers, linters, languages, autocomplete) without explosion. Autocomplete-mode AI looks like that line. Vibe-coding agents would look different. He sees current AI as continuous with the compiler tradition, not a discontinuity.

Mapping against Ray Data Co

Direct alignment. This is, alongside the “What are we scaling?” essay, the strongest external corroboration of the Sanity Check editorial position so far. The “models keep getting more impressive at the rate short-timelines people predict, but more useful at the rate long-timelines people predict” line from the prior Dwarkesh essay rhymes exactly with Karpathy’s “the industry is trying to pretend this is amazing, and it’s not. It’s slop.”

Specific newsletter ammunition:

Sanity Check candidate hooks:

Open follow-ups