Why Andrej Karpathy Says AI Agents Are a Decade From Reality
- Olivia Johnson

- Oct 20
- 8 min read

The technology world is buzzing with anticipation, and many investors and developers have dubbed 2025 "the year of the agent". The promise is tantalizing: intelligent, autonomous AI systems capable of understanding complex goals, breaking them down into steps, and executing them without human intervention. These agents are hailed as the next major leap in artificial intelligence, poised to revolutionize everything from software development to personal productivity.
However, a prominent and deeply respected voice from the heart of the AI revolution is urging for a dose of reality. Andrej Karpathy, a cofounder of OpenAI and a leading figure in the AI community, has delivered a sobering assessment that cuts through the hype. In a recent podcast appearance, he made a bold claim: truly functional AI agents are not just around the corner; they "will take about a decade" to materialize. This analysis unpacks Karpathy's critique, exploring the core limitations he identifies, his vision for a collaborative human-AI future, and why this long-term perspective is crucial for navigating the road ahead.
The Hype vs. Reality: Deconstructing the "Year of the AI Agent"

What Are AI Agents and Why Is Everyone Talking About Them?
Before dissecting the skepticism, it's important to understand what an AI agent is and why the concept has captured the industry's imagination. While definitions can vary, AI agents are generally understood as virtual assistants with a significant degree of autonomy. Unlike current AI chatbots that primarily respond to direct prompts, an agent is designed to complete multi-step tasks on its own. This involves breaking down a high-level problem, formulating a plan, and taking sequential actions to achieve the goal, all without constant user guidance.
The potential applications are vast. Imagine an agent that can book a complex trip with flights, hotels, and transportation based on a simple request. Or a developer agent that can autonomously debug code, write new features, and deploy updates. It is this promise of self-directed, goal-oriented action that has led to a surge in investment and development, fueling the narrative that a new era of AI is imminent.
Andrej Karpathy's Sobering Assessment: "They Just Don't Work"
Despite the widespread enthusiasm, Andrej Karpathy is decidedly unimpressed with the current state of AI agents. His evaluation, shared on the Dwarkesh Podcast, was blunt and unequivocal: "They just don't work". This isn't the critique of an outsider; it's an observation from someone who has been at the forefront of AI development. He argues that the industry's excitement is far outpacing the actual capabilities of today's technology.
Karpathy's critique is not about a lack of effort but a fundamental deficit in the underlying models. He believes the industry is "overshooting the tooling w.r.t. present capability," meaning developers are building complex frameworks and platforms for agents that simply aren't intelligent enough to perform reliably yet. The core systems, he contends, are still "cognitively lacking," which prevents them from moving beyond simple, often brittle, demonstrations.
Core Limitations: Why Today's AI Agents Are "Cognitively Lacking"

Karpathy's assertion that agents don't work is rooted in several specific, deep-seated technical challenges. He identifies a handful of critical areas where current models fall short, making the vision of autonomous operation a distant goal.
The Four Foundational Flaws Karpathy Identified
According to Karpathy, today's agents suffer from a cascade of interconnected failures. He summarized the core issues, stating, "They don't have enough intelligence, they're not multimodal enough, they can't do computer use and all this stuff". Furthermore, he points to a critical missing piece: "They don't have continual learning". This means you can't simply tell an agent something once and expect it to remember and apply that knowledge in the future.
Let's break these down:
Insufficient Intelligence: The base Large Language Models (LLMs) lack the reasoning and planning capabilities to handle the ambiguity and unpredictability of real-world tasks.
Lack of Multimodality: True autonomy requires understanding and interacting with more than just text. Agents need to fluidly process images, videos, and user interfaces to operate effectively, a capability that is still in its infancy.
Poor Computer Use: An agent must be able to reliably navigate websites, use software APIs, and interact with digital tools as a human would. This seemingly simple task is incredibly difficult to execute without errors.
No Continual Learning: Today's models are largely static. They don't learn from their mistakes or remember user preferences from one session to the next, forcing them to start from scratch every time.
The Compounding Error Problem: A Mathematical Barrier
Adding to Karpathy's points, others in the field have highlighted a crippling mathematical reality. Quintin Au of ScaleAI explained that the probabilistic nature of LLMs leads to a "compounding error" problem. Currently, every action an AI takes has a significant chance of error—he estimates this at roughly 20%.
This might not sound catastrophic for a single action, but it becomes devastating over a sequence of steps. As Au pointed out, if an agent needs to perform just five actions to complete a task, the probability of it succeeding at every single step is only about 32%. A task requiring ten steps would have a success rate of just over 10%. This is the stark mathematical reason why agents often fail in complex, real-world scenarios. Until the per-action accuracy of LLMs approaches near-perfection, agents will remain unreliable for anything but the simplest, most constrained tasks.
A New Vision for AI: Collaboration Over Autonomy
Karpathy's critique isn't just about pointing out flaws; it's about redirecting the industry toward what he sees as a more productive and desirable future. He is wary of the prevailing Silicon Valley dream of fully autonomous entities that render human involvement obsolete.
Karpathy's Critique of the "Useless Human" Future
The ultimate vision for many in the agent space is a world where "fully autonomous entities collaborate in parallel to write all the code and humans are useless". Karpathy is explicitly opposed to this goal. He argues that the downside of building agents that make humans useless is precisely that: "humans are then useless". This path, he warns, leads to a world filled with "AI slop"—low-quality, machine-generated content and solutions that degrade our information ecosystem and devalue human skill. He doesn't want to live in that future.
The Ideal Partnership: How Humans and AI Should Work Together
Instead of replacement, Karpathy champions collaboration. His ideal AI is not a black-box automaton but a powerful, transparent partner that augments human capabilities. He envisions an AI that actively works with the user. "I want it to pull the API docs and show me that it used things correctly," he wrote. "I want it to make fewer assumptions and ask/collaborate with me when not sure about something".
In this model, the AI serves as an interactive tool that empowers the user, rather than supplanting them. The goal is for the human to "learn along the way and become better as a programmer, not just get served mountains of code that I'm told works". This human-in-the-loop approach values transparency, verification, and shared learning, creating a positive feedback loop where both human and machine improve together.
The Decade-Long Roadmap: What Needs to Happen Next?
Projecting the Timeline: Why a Decade is a Realistic Estimate
Given the scale of the challenges, Karpathy's projection that it "will take about a decade to work through all of those issues" begins to look less pessimistic and more pragmatic. The four foundational flaws—intelligence, multimodality, computer use, and continual learning—are not minor engineering problems. They are fundamental research hurdles that require significant scientific breakthroughs.
Achieving near-100% reliability on individual actions to solve the compounding error problem is a monumental task. Developing true continual learning, where a model can update its knowledge base safely and effectively without forgetting past information, is one of the holy grails of AI research. These are not problems that can be solved in a single product cycle or with more venture capital alone; they require patient, foundational work over many years.
Positioning Karpathy's View: Skepticism vs. Pessimism
It is crucial to correctly frame Karpathy's position. He is not an AI doomsayer or a skeptic who denies the potential of the technology. He clarifies his own stance by stating that his timelines are "about 5-10X pessimistic" compared to the typical hype found at a San Francisco AI party or on social media.
However, he is also quick to add that his views are "still quite optimistic w.r.t. a rising tide of AI deniers and skeptics". His is a nuanced perspective from an expert in the trenches: he sees the immense potential but also has a clear-eyed view of the immense difficulty of the path to get there. His is a call for patience, intellectual honesty, and a focus on solving the hard problems first.
Conclusion: Navigating the Future of AI Agents with Patience

Andrej Karpathy's assessment serves as a vital course correction for an industry prone to cycles of hype and disillusionment. While the vision of autonomous AI agents is powerful, the technology required to make them a reliable reality is still in its infancy. The path forward is not a short sprint fueled by ambitious tooling but a decade-long marathon focused on solving fundamental challenges in intelligence, interaction, and learning.
By advocating for a collaborative model where AI augments rather than replaces human expertise, Karpathy offers a more sustainable and ultimately more productive vision for the future. His insights remind us that true innovation requires not only ambition but also patience and a deep respect for the complexity of the problems we are trying to solve. The "year of the agent" may indeed come, but it will be the result of a decade of focused, foundational work, not a single year of frantic development.
Frequently Asked Questions (FAQ)
1. What are AI agents, and why are they considered a major innovation?
AI agents are autonomous systems designed to complete complex, multi-step tasks without continuous human input. They are considered a major innovation because, unlike simple chatbots, they can understand a high-level goal, create a plan, and execute actions to achieve it, promising to automate a wide range of digital and cognitive work.
2. According to Andrej Karpathy, what are the biggest problems with current AI agents?
Andrej Karpathy states that current AI agents "just don't work" because they are "cognitively lacking". He identifies four key flaws: they lack sufficient intelligence, are not multimodal enough to process different data types, cannot reliably perform computer-based tasks, and do not possess continual learning to remember new information.
3. How does the "compounding error rate" affect an AI agent's reliability?
The compounding error rate is a critical barrier to agent reliability. As noted by ScaleAI's Quintin Au, if an AI has even a small chance of error on each action (e.g., 20%), the probability of completing a multi-step task successfully drops exponentially with each additional step. For a five-step task, this results in only a 32% chance of success, making agents unreliable for complex workflows.
4. What is Andrej Karpathy's vision for human-AI collaboration?
Karpathy advocates for a future where AI acts as a collaborative partner rather than a replacement for humans. He envisions an AI that shows its work, asks for help when uncertain, and helps the user learn and improve. This model prioritizes transparency and human empowerment over full, unsupervised autonomy.
5. Is Andrej Karpathy an AI skeptic?
No, Karpathy does not consider himself an AI skeptic or denier. He describes his timelines as more pessimistic than the industry hype but still "quite optimistic" compared to true skeptics. His position is that of a pragmatic expert who acknowledges the technology's massive potential but is realistic about the long and difficult road required to achieve it.
6. Why does Karpathy believe it will take a decade for AI agents to be functional?
Karpathy's decade-long timeline is based on the depth of the fundamental challenges that need to be solved. Overcoming issues like insufficient intelligence, the lack of continual learning, and the compounding error problem are not simple engineering fixes but major research hurdles that will require years of dedicated scientific work.
7. What is "AI slop" and how does it relate to autonomous agents?
"AI slop" is a term Karpathy uses to describe the low-quality, often inaccurate or nonsensical content generated by AI. He warns that a future dominated by fully autonomous agents that render humans useless would lead to this "slop" becoming ubiquitous, degrading the quality of our digital information and creative work.


