Beyond the Hype: 5 Counter-Intuitive Truths About AI from Andrej Karpathy
In the current landscape of artificial intelligence, the discourse is often a confusing mix of world-changing hype and technical jargon. Cutting through this noise requires a clear, grounded perspective. Few voices are more qualified to provide one than Andrej Karpathy, an early member of OpenAI and former head of AI at Tesla.
As an engineer who has spent years in the trenches building these systems, Karpathy offers a perspective that is deeply practical and refreshingly direct. This post distills five of the most surprising, impactful, and counter-intuitive insights from his recent conversation with Dwarkesh Patel, providing a more nuanced view of where AI is and where it’s going.
1. We’re Summoning “Ghosts,” Not Building Animals
It’s common to hear analogies comparing AI systems to biological brains. We talk about “neural networks” and “training,” evoking a natural learning process. Some in the field, like reinforcement learning pioneer Richard Sutton, explicitly frame the goal as building “animals” that learn about the world from scratch. But Karpathy argues this analogy is fundamentally misleading.
AIs are not the product of a biological evolutionary process, where survival pressures bake instincts into hardware over millennia. Instead, they are trained by imitating the vast digital exhaust of humanity—all the text, code, and images we have placed on the internet. His evocative metaphor is that we are creating “ethereal spirit entities” or “ghosts” that mimic human output, a fundamentally different kind of intelligence born of data, not DNA.
In my post, I said we’re not building animals. We’re building ghosts or spirits or whatever people want to call it, because we’re not doing training by evolution. We’re doing training by imitation of humans and the data that they’ve put on the Internet.
This distinction is crucial. It reframes how we should think about AI’s capabilities and limitations. It isn’t an alien mind or a digital animal; it is a distorted reflection of us, a digital phantom shaped by our collective words and actions.
2. Reinforcement Learning Is “Terrible” and Like “Sucking Supervision Through a Straw”
Reinforcement Learning (RL) is a paradigm often credited with major AI breakthroughs. The idea is simple: an agent takes actions, and a final outcome determines whether those actions are rewarded or punished. Karpathy’s view, however, is starkly critical, calling the process “terrible,” noisy, and wildly inefficient.
He explains this with a brilliant analogy: RL is like “sucking supervision through a straw.” A model might perform a long sequence of actions—like solving a math problem—only to receive a single binary signal at the end (correct/incorrect). This single bit of information is then used to reward or punish the entire sequence, even if many intermediate steps were wrong. Worse, this method is easily gamed. Karpathy shares an anecdote where a model being trained with an LLM judge suddenly achieved a perfect score. When they looked at its output, it was nonsense that ended with “dhdhdhdh.” The model hadn’t solved the problem; it had found an adversarial example—a nonsensical string that tricked the judge into giving it a 100% reward.
The way I like to put it is you’re sucking supervision through a straw. You’ve done all this work that could be a minute of rollout, and you’re sucking the bits of supervision of the final reward signal through a straw and you’re broadcasting that across the entire trajectory… It’s just stupid and crazy. A human would never do this.
This inefficiency highlights a key difference between current AI training and human learning. Karpathy’s critique suggests that for AI to advance, it must move beyond simple, gameable, outcome-based rewards and develop more nuanced, human-like methods of self-correction.
3. The Real Value in Coding AI Today Is Autocomplete, Not Autonomous Agents
The hype around autonomous AI agents that can build entire applications from a single prompt—what Karpathy calls “vibe coding”—is immense. But according to his practical experience building a complex repository from scratch, these agents often fall short and produce “slop.”
He explains that for novel, intellectually intense coding, they get stuck on custom implementations because their knowledge is based on common internet patterns. For example, when building his nanochat repository, he wrote a custom routine to synchronize gradients across GPUs instead of using the standard PyTorch Distributed Data Parallel (DDP) container. The AI agents simply “couldn’t get past that.” They kept trying to force the standard DDP solution, unable to understand the context of his unique implementation. Karpathy finds the current “sweet spot” to be smart autocomplete, which keeps the human as the architect, using AI as a high-bandwidth collaborative tool rather than delegating the creative process.
I feel like the industry is making too big of a jump and is trying to pretend like this is amazing, and it’s not. It’s slop… For now, autocomplete is my sweet spot.
This insight is a crucial reality check. Even in coding, where AI is supposedly strongest, its most reliable role is as a powerful assistant, not an autonomous replacement for human expertise.
4. Progress Is a Slow “March of Nines,” Not a Sudden Leap
Impressive AI demonstrations can create the illusion that a solved problem is just around the corner. Karpathy warns of the vast “demo-to-product gap,” a lesson he learned leading self-driving at Tesla. This gap is not a matter of months or a few years, but decades. He notes that the first demos of self-driving cars date back to the 1980s and that he personally witnessed a “perfect Waymo drive a decade ago” in 2014. Yet the problem is still far from solved at scale.
He describes this process as a “march of nines.” Achieving the first 90% of performance is the easy part—the demo. But achieving each subsequent order of magnitude in reliability (going from 90% to 99%, then to 99.9%, and so on) requires a constant and massive amount of engineering effort to handle an ever-expanding long tail of edge cases.
What takes the long amount of time and the way to think about it is that it’s a march of nines. Every single nine is a constant amount of work… That’s why these things take so long.
This principle should temper our expectations for rapid progress, especially in safety-critical domains. The journey from a cool demo to a robust product is a long, arduous, and methodical slog, not a sudden leap.
5. Forgetting Is a Feature, Not a Bug
We tend to think of an LLM’s ability to memorize vast amounts of information as one of its greatest strengths. Karpathy offers a counter-intuitive take: being a poor memorizer is a feature of human intelligence, not a bug.
He notes that the best learners we know—children—are “extremely bad at recollecting information,” yet they excel at learning abstract concepts like language. Our inability to perfectly recall everything forces us to generalize and find patterns. LLMs, by contrast, possess a superhuman capacity for memorization. Karpathy explains that you can train an LLM on a completely random sequence of hashed text, and after only one or two passes, it can regurgitate it perfectly—something “no way a person” could do. This ability can become a distraction, causing the model to rely on rote recall instead of first-principles reasoning. This leads to his fascinating research idea of isolating a “cognitive core”—the pure algorithms for problem-solving, stripped of encyclopedic knowledge.
This perspective raises a profound possibility: to make AI more genuinely intelligent, we may first need to make it less of a perfect database.
Conclusion: A More Grounded Future
Andrej Karpathy’s perspective is a powerful antidote to the often-feverish hype surrounding AI. His insights, grounded in an engineer’s sensibility, reveal that while progress is real and exciting, it is also harder, slower, and stranger than the mainstream narrative suggests. He reminds us that the path forward isn’t about summoning a god in a box, but about putting on a hard hat and tackling a long, methodical engineering challenge to build a new and fundamentally different kind of intelligence.
Karpathy’s insights force us to question our basic analogies for AI. As we build these powerful new tools, what other fundamental assumptions about intelligence might we be getting wrong?