Claude Code Review: Anthropic Deploys Multi-Agent Teams to Catch Bugs in Pull Requests

Noqta Team
By Noqta Team ·

Loading the Text to Speech Audio Player...

Anthropic has launched Code Review for Claude Code, a multi-agent system that dispatches teams of AI agents to review pull requests in depth. Available now in research preview for Team and Enterprise plans, the tool is modeled on Anthropic's own internal review process — and it's already changing how code gets shipped.

The Problem: Code Output Is Up, Review Quality Is Down

Code output per engineer at Anthropic has grown 200% in the last year, thanks largely to AI coding tools like Claude Code, Cursor, and others. But that productivity surge has created a bottleneck: code review. Developers are stretched thin, and many PRs get quick skims rather than deep reads.

This is the exact pattern we see across the industry. More code ships faster, but quality assurance hasn't kept pace. Claude Code Review is Anthropic's answer to that gap.

How It Works: Agent Teams, Not Single Passes

When a pull request is opened, Code Review dispatches a team of agents that work in parallel. Here's the process:

  1. Multiple agents analyze the PR simultaneously, each looking for different bug categories
  2. Bugs are verified to filter out false positives
  3. Findings are ranked by severity
  4. A single overview comment lands on the PR, plus in-line comments for specific issues

Reviews scale with complexity. Large, complex PRs get more agents and deeper analysis. Trivial changes get a lightweight pass. Average review time: ~20 minutes.

🔍 Need professional code auditing for your team? Noqta's Vibe Coding Audit & QA service combines AI-powered analysis with human expertise to ensure your codebase stays clean.

The Numbers: Internal Results at Anthropic

Anthropic has been running Code Review internally for months. The results are striking:

  • Before Code Review: 16% of PRs received substantive review comments
  • After Code Review: 54% of PRs receive substantive comments — a 3.4x increase
  • Large PRs (1,000+ lines): 84% get findings, averaging 7.5 issues per review
  • Small PRs (under 50 lines): 31% get findings, averaging 0.5 issues
  • False positive rate: Less than 1% of findings marked incorrect by engineers

In one case, a seemingly routine one-line change to a production service was flagged as critical — it would have broken authentication for the entire service. The engineer admitted they wouldn't have caught it on their own.

Real-World Catch: TrueNAS Encryption Bug

Early access customers are already seeing results. On a ZFS encryption refactor in TrueNAS's open-source middleware, Code Review found a pre-existing bug in adjacent code: a type mismatch that was silently wiping the encryption key cache on every sync. It wasn't even in the PR's changeset — it was a latent issue the PR happened to touch.

This is the kind of deep, contextual analysis that separates multi-agent review from simple linting.

Pricing and Controls

Code Review is positioned as a premium, depth-first tool — not a cheap linter replacement:

  • Average cost: $15–25 per review, scaling with PR size
  • Monthly organization caps to control spend
  • Repository-level control — enable only where you need it
  • Analytics dashboard for tracking review costs and acceptance rates

For teams wanting a lighter option, the Claude Code GitHub Action remains open source and free.

What This Means for Development Teams

Claude Code Review fits into a growing trend: AI is moving from code generation to code governance. We covered Cursor's Automations launch last week, which focuses on always-on agents for security and incident response. Anthropic's approach is different — focused specifically on deep PR review.

The pattern is clear: as AI coding tools generate more code, the industry needs equally capable tools to verify it. This is why comparing these tools matters more than ever — the review layer is becoming just as important as the generation layer.

For organizations building AI agent workflows, Code Review also demonstrates a compelling multi-agent architecture pattern: dispatch specialized agents in parallel, verify results, then synthesize into a single actionable output.

💡 Building with AI coding tools and need quality assurance? Noqta helps teams implement AI-assisted development workflows with proper review and governance built in.

How to Get Started

  1. Admins: Enable Code Review in Claude Code settings, install the GitHub App, select repositories
  2. Developers: Once enabled, reviews run automatically on new PRs — no configuration needed
  3. Docs: Full documentation

FAQ

Does Claude Code Review replace human reviewers?

No. It won't approve PRs — that remains a human decision. It catches issues so reviewers can focus on architecture and design rather than bug hunting.

How much does it cost?

Reviews average $15–25, scaling with PR size. Organizations can set monthly caps. The free Claude Code GitHub Action is available for lighter reviews.

Which platforms are supported?

Currently GitHub only, available for Team and Enterprise Claude Code plans.

How accurate are the findings?

Less than 1% of findings are marked incorrect by Anthropic's internal engineers. Large PRs average 7.5 issues found.

How does it compare to Cursor Automations?

Cursor Automations focuses on always-on agents for security and incident response. Code Review focuses specifically on deep, multi-agent pull request analysis. Different tools for different problems.


Want to read more news? Check out our latest news article on Apple Delays Gemini-Powered Siri Features Beyond iOS 26.4 Due to Technical Issues.

Discuss Your Project with Us

We're here to help with your web development needs. Schedule a call to discuss your project and how we can assist you.

Let's find the best solutions for your needs.