1.6M Git Events Show AI Code Needs More QA

Fazm Team··2 min read

1.6M Git Events Show AI Code Needs More QA

Analysis of large-scale git activity is revealing an uncomfortable trend: AI-generated code is scaling faster than the QA processes designed to catch its mistakes. The code ships faster, but the review quality has not kept up.

The Volume Problem

When a single developer can generate 10x more code with AI assistance, the review bottleneck shifts. Previously, the limiting factor was writing speed. Now it is review capacity. A team of five developers using AI coding tools can produce the output of 20-30 developers, but they still have five humans available for code review.

The math does not work. If each developer generates 500 lines of code per day instead of 100, and review throughput stays flat, the review backlog grows every single day. Teams respond by either letting reviews slip or rubber-stamping changes. Both options increase bug rates.

What the Data Shows

Projects with high AI-generated code ratios show three patterns. First, PR sizes increase - instead of focused 50-line changes, PRs regularly hit 200-500 lines because AI generates full implementations rather than incremental changes. Larger PRs get worse reviews because reviewer fatigue sets in.

Second, test coverage does not scale proportionally with code volume. AI generates implementation code faster than test code, and developers often skip writing tests for AI-generated code they assume is correct.

Third, rollback frequency increases. More code ships, more bugs ship, and more emergency reverts happen. The net velocity - code that stays deployed - is lower than the gross velocity suggests.

Scaling Review

The practical solution is not reviewing every line. It is reviewing the right lines. Focus human attention on architecture decisions, security boundaries, and business logic. Let automated tools handle style, formatting, and basic correctness.

Set up AI-powered review as a first pass - let one model review another model's output. This catches the obvious issues. Human reviewers then focus on the things AI reviewers miss: whether the code solves the right problem, whether the approach fits the architecture, and whether the edge cases are handled.

Fazm is an open source macOS AI agent. Open source on GitHub.

More on This Topic

Related Posts