Picture a developer at a Fortune 500 company, shipping features at 1.8x her normal pace, feeling like a superhero. Then picture her 3 months later, buried in a refactoring nightmare that GitHub's own data says is 31% worse than it used to be. That is the AI coding story nobody is putting on a conference slide.
I am not a software engineer. I am the guy who writes about tech for people who are not software engineers. But I have spent the last week talking to developers, reading the actual research, and watching the discourse, and here is my honest read: AI coding tools make you faster the way a credit card makes you richer. The number goes up immediately. The reckoning comes later.
The Speed Is Not a Lie
GitHub's Octoverse report from March 20 analyzed 1.5 million Copilot users and found a 55% faster task completion rate. McKinsey surveyed 2,000 engineers at Fortune 500 firms and found a 42% productivity gain in code writing speed. Those numbers are not made up. Andrej Karpathy, who built some of this stuff at OpenAI, called AI code "prolific but sloppy" on the Lex Fridman podcast last week. That is a pretty good one-sentence summary of everything I found.
The problem is what happens after the code ships. GitHub's same report found bug rates in AI-assisted codebases are up 12%. Microsoft's CodeGenBench 2.0, published April 5, found GPT-4o code fails integration tests 25% more often than human-written code. Anthropic's internal audit of Claude 3.5 Sonnet found refactoring needs rose 31% post-deployment. And on March 15, Copilot-generated code caused a 4-hour Uber outage that hit 10 million rides. That is not a benchmark. That is a real thing that happened to real people trying to get home.
You Are Now an Editor, Not a Writer
Casey Newton at Platformer put it cleanly: McKinsey's data shows AI turns engineers into editors, not creators. Faster output, but at the cost of deeper understanding. I think that framing is right, and I think it matters more than the speed stats.
Here is the tension I keep running into: the speed gains are real and the sloppiness is also real, and they are not canceling each other out. They are compounding. You ship faster, so you ship more. More code means more surface area for bugs. More bugs means more debugging time, which McKinsey clocked at 19% higher. The code acceptance rate for AI suggestions dropped from 35% to 28% in a single year, which means engineers are already getting pickier. But pickier is not the same as careful.
Fair point to the optimists: Nat Friedman, former GitHub CEO, argued that engineers who review AI code end up 2x more productive long-term. That is probably true for senior engineers who already know what bad code looks like. It is not true for the junior dev who has never seen a codebase without AI assistance and does not know what they are accepting.
That is the specific group I am worried about. Not abstract "engineers." The person two years into their career who has never had to write a function from scratch, who accepts 28% of suggestions without fully understanding the other 72% they rejected. That person is accumulating debt they cannot see yet.
So here is my actual take: use the tools, they are genuinely good, but engineering teams need mandatory code review training that is specifically about AI output, not just general best practices. The speed gains are worth keeping. The 31% refactoring spike is not inevitable. It is what happens when nobody builds the guardrails.
The Uber outage did not have to happen. Someone just had to read the code.