Karpathy on AI as Ghost Intelligence and Capability Spikes
Press Space to continue
Karpathy’s 2025 retrospective is the clearest articulation I’ve seen of what foundational AI labs are actually building. We’re not “evolving animals,” we’re “summoning ghosts.” LLMs have completely different optimization pressures than biological intelligence. Humans evolved for tribal survival. LLMs optimize for imitating text, solving puzzles, and winning upvotes on LM Arena. Different pressures, different shapes in the intelligence space. This framing finally explains what confuses everyone about AI capability. GPT-5 aces the bar exam but gets tricked by simple jailbreaks. Claude writes PhD-level philosophy but hallucinates citations. Gemini solves competition math that stumps IMO medalists but fumbles spatial reasoning. Capability spikes near verifiable domains where RLVR concentrates optimization pressure. Everywhere else, you get a different entity entirely. I’ve been thinking about what this means for AI product builders. The teams struggling with AI deployment are treating capability as uniform. They ask “can AI do this task?” and expect a yes/no answer. But ghost intelligence doesn’t work that way. The teams winning are asking a different question: “Does this task live near a verifiable domain?” If yes, the ghost might be superhuman. Build for autonomy. If no, the ghost needs guardrails. Build for human-in-the-loop. This is why Cursor works. This is why Claude Code runs on localhost instead of the cloud. The best AI products in 2025 mapped the jagged edges and designed around them. The companies that internalize Karpathy’s ghost framing will build better products than the ones still thinking in terms of “smarter or dumber than humans.” There’s no single axis. Just different shapes.