The AI Productivity Trap: Why Shipping Faster Isn’t the Same as Building Better
New data shows AI-assisted developers generate 9.4x more code churn than their peers. The tools aren’t the problem. Mistaking output for progress is.
There’s a number making the rounds in engineering circles that should make every founder uncomfortable: AI-assisted developers produce 9.4x more code churn than their non-AI counterparts.
Churn, for the uninitiated, is when code gets written and then rewritten, deleted, or replaced before it ships. It’s the engineering equivalent of spinning your wheels. The car moves, the odometer ticks up, but you’re not going anywhere meaningful.
The finding comes from GitClear’s analysis of over 211 million lines of code. It did not bury this result. And yet the dominant narrative in tech remains: AI makes developers faster, therefore AI makes teams better. That leap deserves more scrutiny than it’s getting.
Output Is Not Progress
Founders are wired to celebrate velocity. Ship fast. Iterate. Get things out the door. That instinct served us well in an era where the constraint was mostly human throughput: hours in the day, fingers on keyboards, brains on problems. More code per developer meant more product per quarter.
AI tools did not change the game. They changed the unit of scarcity. Time is no longer the bottleneck. Judgment is.
When a developer can generate a working implementation in seconds, the question stops being “can we build this?” and starts being “should we build this, and should we build it this way?” Those are harder questions. They require taste, context, and a willingness to say no to things that technically work but strategically do not.
The 9.4x churn stat is evidence that most teams have not made that mental shift. They are using AI to go faster in the same direction they were already going. Sometimes that direction is wrong. Sometimes the implementation is sloppy. Either way, you end up rewriting it.
Why AI Makes This Worse, Not Better
Here is the uncomfortable part: the tools are working exactly as designed. They are optimized to produce plausible output quickly. “Plausible” and “correct” are not the same thing. Neither are “quickly” and “efficiently.”
When a developer writes code by hand, there is a natural forcing function. The effort of writing creates a feedback loop. You think through edge cases because you have to touch every line. You notice that the function you are writing already exists in a slightly different form three files over. You feel the drag of complexity accumulating.
AI removes that friction. And friction, it turns out, was doing a lot of work.
The result is what the data shows: more code produced, more code discarded. The team feels productive because the commit history is full. The codebase feels productive because features are shipping. But underneath, technical debt compounds faster than anyone realizes, because no one is doing the slow, uncomfortable work of actually understanding what is being built.
What the Founder Should Actually Be Measuring
If your engineering team has adopted AI coding tools (which they almost certainly have), the metric you should be watching is not velocity. It is durability.
How much of what ships in month one is still running unchanged in month three? How often are your engineers rewriting the same module? What is the ratio of net new functionality to refactors and bug fixes? These numbers tell a different story than story points completed or PRs merged.
Durability is a proxy for judgment. When code survives, it usually means someone thought carefully about what they were building before they built it. When code churns, it usually means someone built the first thing that worked rather than the right thing.
This is not an argument against AI tools. It is an argument for using them deliberately. The teams that will win with AI are not the ones who use it most. They are the ones who use it on the right problems, with the right guardrails, and with enough architectural clarity that AI-generated code has somewhere sensible to land.
The Process Problem Underneath
High churn is almost never a tools problem. It is a process problem wearing a tools costume.
Teams that churn a lot with AI assistance were usually churning before, just more slowly. The AI amplifies whatever was already there: unclear requirements, weak code review, insufficient architecture discussion before implementation begins. Ship-first-think-later cultures do not get fixed by better tooling. They get exposed by it.
If your team is shipping faster but rewriting constantly, the diagnosis is almost certainly upstream. Requirements are fuzzy. Design sessions are short or skipped. Engineers are starting to code before they understand the problem. Those are organizational issues, not technical ones.
The fix is not to slow down AI usage. It is to front-load the thinking that AI cannot do for you: what problem are we actually solving, who is it for, what does done look like, and what are the constraints we are building inside of? Get those answers before the AI writes a single line, and churn drops. Not because you are using the tool differently, but because you gave the tool a better problem to work on.
Speed Is a Feature, Not a Strategy
The founders who use AI most effectively treat it like a very fast junior developer. You would not hand a junior developer a vague brief and tell them to ship. You would give them a clear spec, a defined scope, and checkpoints to verify they are on track. AI deserves the same discipline.
What gets celebrated in most dev teams right now is throughput: how many tickets closed, how many features shipped, how fast the sprint burned down. Those are fine metrics for the short term. They are terrible proxies for whether you are building something that will last.
The AI coding productivity trap is not that the tools are bad. It is that they make it very easy to feel like you are making progress when you are actually generating sophisticated-looking work that will need to be undone. Speed becomes the story you tell yourself to avoid the harder conversation about whether you are building the right things, in the right way, for the right reasons.
There is a version of building fast that is also building well. That version requires judgment, clarity, and the discipline to let AI accelerate the execution of good decisions rather than the iteration around bad ones. Most teams are not there yet. The data makes that pretty clear.