What Will Kill Claude Code
On the inevitable collapse of agent coding tools under their own complexity
Markus Hav
Lead Researcher, Agents · March 22, 2026
Abstract
After extensive testing of Opus 4.6 in Claude Code, a pattern has become clear: agent coding tools are slowly collapsing under their own complexity. Each new model generation should enable simplification, but legacy engineering intuition pushes toward more layers instead. This essay argues that Claude Code, in its current architectural form, has roughly 3–6 months of relevance left—and identifies two successor architectures that could replace it.
The Magic Is Real—But Not for Everyone
Opus 4.6 in Claude Code feels magical. For newcomers to AI-assisted coding and for complex, legacy codebases unfamiliar to agents, the experience is genuinely transformative. The new model enables agentification of previously intractable code at a level we haven't seen before.
But there is an enormous BUT.
If your codebase is AI-native—already optimized to generate well since, say, Sonnet 3.7—the experience is honestly wasteful. Once again, we get dramatically more token consumption and latency. A two-minute task where I could hand-craft the perfect context for a one-shot solution spawns five explore agents despite the entire plan already being laid out. By the time execution should begin, the context window is full for the second time, bloated with outputs from dozens of agents it launched.
The background agent orchestration makes this worse. When execution cannot continue until all agents complete—which is often the case—keeping Claude "warm" produces an excruciating stream of filler: "let's wait, the architect agent isn't done yet, oh that one is, but we're still waiting for these." Pure token waste, pure frustration.
The Complexity Trap
Zoom out, and a pattern emerges that has been visible for months: Claude Code is slowly collapsing under its own complexity. And it is not the first agent coding tool to do so.
The Complexity Trap
In the AI era, every new model is an opportunity to simplify. If your inherited intuition from traditional software engineering tells you to add complexity, you are walking slowly toward your own death. Every layer of cleverness you stack on top is unnecessary legacy for the next, genuinely smarter model.
The most telling symptom: Anthropic is now optimizing the language model itself to handle the tool's complexity. This is deeply backwards. They are training the model to navigate the scaffolding, rather than removing the scaffolding the model no longer needs.
I have not once, in the past eighteen months, regretted throwing away a codebase and starting fresh.
A Generous Timeline: 3–6 Months
In fairness: it is entirely possible—and even consistent with accepted AI scaling theories—that Anthropic and Claude could identify and fix these problems themselves. But if we set aside singularity speculation and look at the facts, what remains is this: we have a year-old product from a company whose core competency is building language models, not developer tools. I have no strong evidence they can solve what is fundamentally a product architecture problem.
We know how lost language models still are when it comes to agent orchestration.
To be clear: I continue as a satisfied Claude Code user and recommend it without reservation. But in its current form, I give it 3–6 months.
What Comes Next
Two architectural shifts could replace the current generation of agent coding tools:
Swarm + One-Shot
A thousand lightweight agents gather full codebase context in seconds. A super model then one-shots the entire solution. Execution time collapses to minutes for any problem size.
Fully Persistent Context
An agent that grows and learns implicitly with every interaction. No CLAUDE.md files, no context window limits, no compression, no clearing. The conversation history concept itself is eliminated.
1. Massive-Scale Swarm Orchestration
Forget the human dev-team analogies that current tools are built around. Instead: a thousand lightweight agents that collect context from the entire codebase in tens of seconds. With that efficient panoramic view, a super model (Opus, Gemini Pro) one-shots the complete solution.
Execution time would effectively flatten to a few minutes regardless of problem complexity. The agent could even mathematically derive task complexity from the full-codebase context and scale its approach accordingly.
2. Fully Persistent Context
What if, instead of starting fresh each session, your agent grew and implicitly learned from every single interaction? You could forget about CLAUDE.md files, context window limitations, compression, clearing, and the many other oddities of current Claude Code life.
This fundamentally changes what an "agent" even means—and where the human sits in the process. Because it redefines these concepts so radically, it may be difficult for builders of "traditional" agents, including the Claude Code team, to make this leap. Exploding the concept of conversation history requires abandoning deeply held assumptions.
Both architectures could theoretically be built into Claude Code. But each one changes the agent concept and the human's role so fundamentally that it may effectively require a new product.
When the settings, flags, and effort knobs in your tool keep multiplying, an AI-native thinker should recognize the signal: there is a meta-level gain waiting to be captured. Complexity is the symptom. Simplification is the product.
The tools that survive will be the ones brave enough to subtract.
About the Author
Markus Hav
Markus Hav is Lead Researcher for Agents at Benque Max AI Lab in Finland, where he focuses on advancing autonomous AI systems and agent architectures. His work explores the boundaries between programmed behavior and emergent intelligence in AI agents. He also serves as Head of AI Automation at Hoxhunt, applying cutting-edge agent research to real-world automation challenges.