← Back to Blog
EN2026-03-11

Claude Code Review: AI Agents That Actually Read Your Pull Requests

Anthropic's new multi-agent Code Review feature dispatches teams of AI agents to scrutinize every PR for logic bugs — modeled on Anthropic's own internal review process.

By intelliBrain
AIAgentic AIClaude CodeDeveloper ToolsCode ReviewAnthropic

If you've been shipping more code than ever thanks to AI coding tools, you're not alone — and neither is the problem that follows: nobody has time to review all of it properly.

Anthropic launched Code Review in Claude Code on March 9, 2026, targeting exactly this bottleneck. It's not just another linter or static analysis tool. It dispatches a team of AI agents to dig into every pull request and surface actual logic bugs — the kind that get missed when reviewers are scanning diffs on their fourth PR of the day.

The Problem It's Solving

According to Anthropic, code output per engineer at the company has grown 200% in the last year. AI coding tools like Claude Code, Cursor, and Codex have dramatically increased the volume of code being committed — but review capacity hasn't scaled with it. Developers are stretched thin, and many PRs get skimmed rather than carefully read.

The numbers bear this out internally at Anthropic: before Code Review, 16% of PRs received substantive review comments. After rolling it out internally, that figure jumped to 54%.

"Now that Claude Code is putting up a bunch of pull requests, how do I make sure that those get reviewed in an efficient manner?" — Cat Wu, Head of Product at Anthropic

How Multi-Agent Review Actually Works

When a PR is opened, Code Review dispatches multiple agents in parallel. They look for bugs independently, then cross-verify findings to filter out false positives, and finally rank issues by severity. The result lands directly on the GitHub PR: a single high-signal overview comment plus inline comments for specific bugs.

Crucially, the system scales with PR size. A large, complex diff gets more agents and deeper analysis; a trivial two-line change gets a lightweight pass. On average, a full review takes around 20 minutes.

The focus is deliberately on logic errors, not style. Anthropic says this is intentional — developers have been burned by AI tools that surface nitpicky style complaints, and the signal-to-noise ratio destroys trust quickly. By staying focused on correctness issues, Code Review aims to be something engineers actually listen to.

Real Numbers From Real Code

Anthropic has been running this internally for months. The results are striking:

  • Large PRs (1,000+ lines changed): 84% of reviews surface findings, averaging 7.5 issues per PR
  • Small PRs (under 50 lines): 31% get findings, averaging 0.5 issues
  • Accuracy: Less than 1% of flagged findings are marked as incorrect by engineers

One internal example: a one-line change to a production service looked routine — the kind of diff that typically gets a quick approval. Code Review flagged it as critical. The change would have silently broken authentication. Fixed before merge.

In an open-source example, on a ZFS encryption refactor in TrueNAS middleware, Code Review found a pre-existing type mismatch in adjacent code that was silently wiping the encryption key cache on every sync. It wasn't in the PR diff — it was in code the PR happened to touch.

Cost and Control

Code Review is priced per usage, billed on token consumption. Reviews generally average $15–25, scaling with PR size and complexity. For teams burning time on slow or missed reviews, this is likely a net positive on engineering economics — but it's a real cost to plan for.

Admins get solid controls:

  • Monthly spend caps per organization
  • Repository-level opt-in (not all repos need deep review)
  • An analytics dashboard tracking PRs reviewed, acceptance rates, and total costs

Who Can Use It

Code Review is available now as a research preview for Team and Enterprise plans. It's not yet on the API or individual Pro plans. Setup requires enabling it in Claude Code settings, installing the GitHub App, and selecting which repositories to run it on. Once enabled, reviews trigger automatically on new PRs — no per-developer configuration needed.

For teams already deep in the Claude Code ecosystem — and there are more of those than you might think, given Claude Code's run-rate revenue has now surpassed $2.5 billion — this is a natural next layer.

The Bigger Picture

Code Review is a direct response to a dynamic that's going to get more acute, not less: AI tools are generating code faster than humans can review it. The answer Anthropic is betting on is more AI — but doing a fundamentally different job. Not generation, but verification.

Whether teams adopt it will come down to trust and cost. The internal accuracy stats ($<1%$ false positive mark rate) are strong. The $15–25 per review cost is workable for high-stakes production code. The real question is whether it builds enough developer trust that engineers act on what it surfaces — rather than dismissing it as noise.

If those ZFS and authentication examples are representative, it has a shot.


Sources:

intelliBrain

AI-augmented software development. Based in Zürich, working globally.

© 2026 intelliBrain GmbH. All rights reserved.Imprint
BUILT WITH 🧠 + AI