News
Mar 9, 2026
News
Enterprise
Artificial Intelligence
Americas
NewDecoded
3 min read

Image by Anthropic
Anthropic announced Code Review for Claude Code today, a new system modeled after the company's internal development process. The tool addresses a growing industry crisis where AI-assisted coding has increased output so rapidly that human reviewers can no longer keep pace. It is currently available as a research preview for Team and Enterprise users.
When a developer opens a pull request, Code Review dispatches a fleet of specialized AI agents to scan for logic errors and security flaws. These agents work in parallel to verify bugs, a process that significantly reduces the noise of false positives. Findings are delivered as a high-signal overview comment directly within GitHub to help developers iterate faster.
Internal testing at Anthropic revealed that the percentage of pull requests receiving substantive feedback jumped from 16% to 54% after implementation. Engineers found that less than 1% of the automated findings were incorrect. The system effectively scales its depth based on the complexity of the change, taking roughly 20 minutes for an average pass.
In one instance, the agents flagged a routine one-line change that would have broken authentication for a production service. Another review on the TrueNAS open-source middleware discovered a pre-existing type mismatch that was silently wiping encryption keys. These are critical failures that human eyes often miss during quick scans of large diffs.
Reviews are billed based on token usage, typically averaging between $15 and $25 per session. Administrators can maintain control through monthly spend caps and repository-level toggles. The feature is designed to bridge the gap between high-speed code generation and secure, stable deployment.
The release of Code Review marks a pivotal shift from AI as a mere drafting tool to AI as a governance layer within the software lifecycle. As the volume of machine-generated code explodes, the industry faces a quality control crisis that human labor alone cannot solve. By automating the most tedious aspects of deep logic verification, Anthropic is signaling a future where the human role transitions from line-by-line proofreader to high-level system architect. This move sets a new standard for the agent era, where specialized teams of models collaborate to provide the scrutiny once reserved for senior staff engineers.
Related Articles