zinch

BLUEPRINT · ENGINEERING · CROSS-VERTICAL

Engineering Code Review

Catch defects before review, draft fixes, ship faster with audit.

Catch defects before review, draft fixes, ship faster with audit. Named scope, named timeline, named stack — ADK · Memory Bank · Evals · 3 weeks.

Cinematic still-life of an engineer workstation with an open pull-request diff and a sapphire-blue glass loupe magnifying the changed lines, representing the Engineering Code Review blueprint.

Design ceiling

Routine review feedback drafted before the human reviewer reads the diff.

The agent system is designed to read the PR against the repo conventions, run the static checks, and draft the routine feedback as inline comments — so the senior reviewer opens the PR with the style, test, and security signals already on the page and spends the time on architecture and intent.

The problem

Senior reviewers are not paid to chase missing tests.

A platform team running disciplined code review is paying its most expensive engineers to flag the same patterns on every PR — a missing test, a swallowed exception, a style drift, a known anti-pattern from the repo conventions doc nobody re-reads. The signal is correct and the work is repetitive, and the reviewer who could be sharpening architecture is sharpening a checklist.

The job is not to replace the human reviewer. The job is to draft the routine feedback first — inline comments on the diff, fix suggestions where the fix is mechanical, the static-analysis output already triaged — so the reviewer arrives to read intent and design, not to point at line 42 again. One agent, Memory Bank carrying the repo conventions across the review, eval set tuned against the team's own historical PRs.

Agent architecture

Four pillars, this blueprint’s stack.

The platform’s four pillars, mapped to the components this agent system actually exercises.

  • Still-life of a graphite pencil and a sapphire-ink architectural schematic on cream paper, representing the Build pillar of Gemini Enterprise Agent Platform.

    Build

    One agent. Reads the diff, applies the conventions, drafts the comments.

    ADK · Agent Studio · Model Garden

  • Overhead photograph of a hexagonal concrete tessellation with sapphire-inset stones, representing the Scale pillar of Gemini Enterprise Agent Platform.

    Scale

    Repo conventions and prior-PR context live in Memory Bank across the review.

    Agent Runtime · Memory Bank

  • Still-life of a precision steel hinge resting on a navy leather ledger, representing the Govern pillar of Gemini Enterprise Agent Platform.

    Govern

    Source code stays in your repo. Every comment posted under a registered identity.

    Agent Registry · Model Armor · Gateway

  • Still-life of brass vernier calipers measuring a row of brass machined cylinders on cream paper, representing the Optimize pillar of Gemini Enterprise Agent Platform.

    Optimize

    Eval set built from your team's historical PRs and reviewer accept rates.

    Evals · Observability · Agent Analytics

Engagement · 3 weeks

Week by week.

Fixed scope, fixed price, fixed timeline. Here is what happens when.

  1. Week 1

    Discovery and reviewer-agent spec.

    Walk recent PRs with the platform lead and a senior reviewer. Decide which signals the agent surfaces and which it suppresses. Capture the repo conventions doc as the agent's reference state.

  2. Week 2

    Reviewer agent and eval pass.

    Build the ADK reviewer agent against your repo. First eval pass on a sampled set of historical PRs against the reviewer accept-or-reject signal. Tighten prompts and tool calls against the eval results.

  3. Week 3

    PR pipeline wiring and shadow run.

    Wire the agent into the PR pipeline. Shadow-run on real PRs for the week with reviewer feedback flowing back into the eval set. Hand off the repo, the eval harness, and the runbook to the platform lead.

What it looks like in code

ADK Python — the agent definition.

The actual shape of the code your team owns at engagement end. Real ADK, real tools, real instruction copy.

agents/code_reviewer/agent.py

python

from google.adk.agents import LlmAgentfrom google.adk.tools import FunctionToolfrom .tools import (    fetch_pr_diff,    fetch_repo_conventions,    run_static_analysis,    post_review_comment,)reviewer = LlmAgent(    name="code_reviewer",    model="gemini-2.0-pro",    instruction=(        "You draft the routine review feedback on a pull request: missing "        "tests, swallowed exceptions, style drift against the repo "        "conventions, static-analysis findings worth surfacing. Post "        "comments inline on the diff. Suppress signals the conventions "        "doc explicitly accepts. Architecture and intent are the human "        "reviewer's call, not yours."    ),    tools=[        FunctionTool(fetch_pr_diff),        FunctionTool(fetch_repo_conventions),        FunctionTool(run_static_analysis),        FunctionTool(post_review_comment),    ],)

What you walk away with

Code your team owns. Governance baked in.

Every blueprint hands the engineering team a deployed agent and the artefacts to run it themselves. No black box, no lock-in.

  • One ADK reviewer agent deployed on Agent Runtime, in your Google Cloud project.
  • The Git repo your team owns, with the agent code, tools, prompts, and tests.
  • CI integration into your PR pipeline (GitHub Actions, GitLab CI, or your existing runner).
  • Eval harness built from your team's historical PRs with the reviewer accept-or-reject signal as ground truth.
  • Repo-conventions reference doc the agent reads from, version-controlled alongside the agent.
  • Agent Analytics dashboard plus an on-call runbook for the platform lead.

Book this blueprint as a pilot.

Two weeks. Named scope. Working agent on Agent Runtime at the end.

  • Code

    Lives in your Git org, owned from commit one.

  • Governance

    Model Armor and Agent Registry on day one.

  • Speed

    Two weeks to a runnable pilot. Eight to production.

Not ready to talk? Take the 4-min readiness assessment