← All articles Reality Check

The 90% AI-coded myth — what 'built with AI' actually means for production

The 90% AI-coded myth — what 'built with AI' actually means for production

The 90% AI-coded myth — what 'built with AI' actually means for production

A growing number of companies proudly announce that their products are "90-95% coded with AI." This is presented as a badge of honor — proof of innovation, efficiency, and forward thinking. Conference speakers cite these numbers to gasps of admiration. Investors hear them as signals of lean operations.

But in production environments, this claim raises a different set of questions. Not "how impressive" but "how sustainable." Not "how fast" but "who owns this when it breaks."

The claim sounds impressive

The appeal is obvious. If AI wrote 90% of the code, development was faster. Fewer developers were needed. The product reached market sooner. For audiences unfamiliar with production software, these numbers suggest a revolution in efficiency.

And for certain contexts — prototypes, proof-of-concepts, internal tools with limited lifespans — a high AI-generation ratio can be entirely appropriate. The problems emerge when the same numbers are celebrated for production systems that handle real users, real data, and real consequences.

The questions nobody asks

When someone claims 90-95% AI-generated code, the follow-up questions matter more than the headline:

Who reviewed it? Every line of AI-generated code requires the same scrutiny as human-written code — often more, because AI-generated code can contain subtle errors that look syntactically correct but are logically wrong.

Who understands it? Understanding code means being able to explain what it does, why it does it that way, and what happens when it fails. If the person who prompted the AI cannot answer these questions, the code is an orphan.

Who debugs it at 3am? Production systems fail. When they do, someone must diagnose the problem under pressure, in an unfamiliar codebase, with users waiting. AI-generated code that nobody fully understands becomes exponentially harder to debug.

Who maintains it in six months? Requirements change. Dependencies update. Security patches need applying. Maintenance requires understanding, and understanding requires someone who has read, reviewed, and internalized the code.

"Du äger varje rad output. Och du kan inte äga det du inte förstår."

The compound interest of ignorance

The problem with high AI-generation ratios is not immediate. Week one looks fine. The code works. Features ship. Metrics look good.

The damage compounds over time, following a predictable trajectory that organizations rarely anticipate.

Week 1-2: The honeymoon. Features ship rapidly. Stakeholders are impressed. The team feels productive. Everything works because nothing has been tested by real-world conditions yet.

Week 4-6: The first cracks. A production bug appears. The developer who prompted the AI cannot explain the affected module. Debugging takes three days instead of three hours because nobody understands the code's internal logic. The fix introduces a new bug because the developer modified code without understanding its dependencies.

Week 8-10: The cascade. Multiple team members have generated code that interacts in ways nobody planned. Integration points fail. Error handling is inconsistent because each AI-generated module handled errors differently. The codebase has grown beyond anyone's comprehension.

Week 12+: The reckoning. A critical failure occurs. The team cannot fix it without essentially rewriting the affected components. Management asks how this happened when the AI-generated code "worked perfectly" just weeks ago. The answer: it always had these problems, but nobody understood the code well enough to see them.

This pattern is not theoretical. It plays out in organizations that treat AI code generation as a replacement for understanding rather than a complement to it.

The ownership test

A practical way to evaluate whether "90% AI-coded" is a strength or a liability: the code ownership audit.

For any critical module in the system, find the developer responsible and ask:

  1. Can you explain the architecture of this module without looking at the code?
  2. Can you identify the three most likely failure points?
  3. Could you rewrite the core logic from memory if the code were deleted?
  4. Can you explain every dependency and why it was chosen?
  5. Could you onboard a new hire to this module in under an hour?

If the answer to most of these is no, the code is not owned — it is rented. And rented code has a way of evicting its tenants at the worst possible time.

What "AI-assisted" should actually mean

The problem is not with AI writing code. The problem is with the ratio being celebrated without context. A more honest framework acknowledges that AI assistance exists on a spectrum, and the appropriate level depends on what is being built.

Structured balance (50-70% AI-assisted): The developer defines architecture, sets constraints, and reviews every significant output. AI handles implementation details within a framework the developer understands and controls. This is how most production work should operate — AI accelerates execution while the developer maintains ownership.

Vibe coding (high AI generation): Valid for throwaway prototypes, internal experiments, and proof-of-concepts where the cost of failure is low and the code has a defined short lifespan. Not valid for production systems.

Hardcore planning (lower AI generation): Complex systems, security-critical components, and unfamiliar territories where every line requires deliberate thought. AI assists with specific implementation tasks but the human drives the design.

The teams that work effectively with AI don't celebrate high generation ratios. They celebrate understanding ratios — the percentage of their codebase that the team can explain, debug, and maintain without the AI that generated it.

The organizational responsibility

When organizations encourage or incentivize high AI-generation ratios without corresponding investment in review processes, they are building on a foundation that erodes over time. The short-term gains are real. The long-term costs are also real — and they compound.

Production-first thinking demands a different metric than "how much did AI write." The relevant questions are: how much of the codebase does the team understand? How quickly can they respond to failures? How confidently can they make changes without introducing regressions?

"AI föreslår. Jag bestämmer. Jag pushar. Ordningen är oförhandlingsbar."

A practical framework: the code ownership audit

For organizations that want to move beyond the vanity metric of AI-generation ratio, a quarterly code ownership audit provides a more honest assessment:

  1. Select critical modules — identify the 20% of the codebase that handles 80% of the risk (authentication, payment processing, data handling, core business logic)
  2. Assign ownership — each critical module has a named developer who is responsible for understanding it completely
  3. Test understanding — the owner explains the module to a peer without referencing the code, including failure modes and dependencies
  4. Document gaps — where understanding is insufficient, schedule dedicated review time — not more AI generation, but human comprehension
  5. Track over time — monitor whether ownership is strengthening or degrading as the codebase evolves

This audit does not slow development. It ensures that the development being done is sustainable.

The bottom line

"90% AI-coded" is not inherently good or bad. It is a ratio that requires context. For a weekend prototype, it might be fine. For a production system handling customer data, it is a question that demands follow-up: who owns this code?

The companies that will thrive with AI are not those that maximize their generation ratios. They are the ones that maximize their understanding ratios — using AI to accelerate work they comprehend and control, not to produce output they cannot maintain.

The real measure of AI maturity is not how much code AI wrote. It is how much of that code the team can stand behind.