Skip to content

AI Hype vs Reality

Published:
4 min read

Claude code is not making your product better

Why the most respected engineers are pushing back on the coding agent hype cycle, and what the Claude code argument actually reveals


Something interesting is happening. The loudest voices in the room are telling you that AI coding agents changed everything. The engineers who actually built the things you use every day are telling you something more complicated.

The K-shaped productivity curve

Labor economists are seeing a K-shaped reality: senior engineers using AI are getting meaningfully more productive, while junior engineers are treading water or getting worse.

The popular venture-backed narrative ignores the bottom half of that K, boasting about teams “clearing six years of backlog in a quarter” or building whole backends in three days. But while agentic coding does reduce time-to-PR, lines of code per hour is the wrong metric.

If engineers are more productive, the rate at which product improves per engineer should be going up.

The best product builders are a canary

Recently, respected builders like Dax (OpenCode.ai), Karri Saarinen (Linear), and David Cramer (Sentry) all noticed the same pattern.

Cramer summarized it best after reviewing his GitHub graphs: LLMs remove the barrier to get started, but they create increasingly complex, unmaintainable software. His verdict: it’s mostly bloat. These aren’t luddites—Dax is literally building an AI agent, and Karri built Linear. Yet they are struggling to see their product improvement velocity accelerate.

When you can generate features at the speed of thought, you stop saying no. The cost structure of software development changes. Suddenly, it’s cheaper to ship crud than to think about what matters.

Why isn’t Claude Code in fast take off mode?

Anthropic boasts that Claude Code is completely Claude-coded. If the machine is writing the machine, the rate of product improvement should be a compounding function (y = e^(cx), exponential growth).

If that were true, Anthropic’s 7-month head start should have created an unbridgeable gap between them and competitors like Codex or Cursor. Instead, everyone is still debating which tool is better. The compound advantage isn’t showing up because something else is bottlenecking product quality, and it was never the code.

That’s not how you’d expect exponential improvement to look.

Lines of code is a cost, not a product

The best engineering cultures treat lines of code as a liability on the balance sheet, not an asset. Every line is a surface for bugs; every feature creates dependencies.

Take Linear vs. Jira. Linear has a fraction of the headcount and codebase size, but infinitely better consumer quality. Facebook isn’t constrained by UI code production – it’s constrained by delivering it reliably to billions. The reward function for great software is compression, not production. Coding agents don’t have a theory of the system to make long-term tradeoffs.

The constraint was never “write code faster.” The constraint was always “figure out what’s worth building.”

The real bottleneck: pushing the frontier

Product quality at the frontier isn’t bounded by how fast you write code. It’s bounded by how fast you come up with ideas good enough to push the frontier.

That kind of quality emerges from taste – from the painful, slow decision to build less. A backlog full of CRUD features is exactly what AI accelerates, but shipping faster doesn’t make your product better. It makes it better when you ship something that makes users care more.

The AI productivity paradox

Camrys for everyone, Ferraris for no one

So, is Claude Code worth it? It depends on where you are.


Summary of Substack Claude Code is Not Making Your Product Better

New posts, shipping stories, and nerdy links straight to your inbox.

2× per month, pure signal, zero fluff.


Edit on GitHub