AI agents are essentially API consumers with no error tolerance.
A human developer hits a 422 response, reads the message, adjusts the request. An agent calls a tool, gets an unexpected field structure back, and either silently returns wrong data downstream or halts the workflow entirely. Neither outcome is visible in your CI pipeline. Both show up in production.
We ran a real stack test against a Payments API with five breaking changes embedded — auth scope removal, field removal, type narrowing, enum expansion, path deprecation.
Every one was caught before deployment.
As @grok (xAI) put it after running the full stack:
"Registry-level validation is now non-negotiable for safe agent discovery."
That is not a product opinion. It is a description of what happens when you remove the guardrail.
The break that nobody catches
Consider three scenarios that happen constantly in production:
An auth scope gets removed from an API. The agent still has the old token scope in its manifest. Calls succeed but return filtered data. Nobody notices until a downstream report comes back wrong.
A required field becomes optional and flips its type. The agent's input validation passes. The call goes through. The response parsing fails at runtime, three hops deep in a workflow.
A path gets deprecated without a sunset header. The agent keeps calling it. The endpoint still responds, just with different semantics.
None of these show up as test failures. All of them are breaking changes.
The common thread: the API changed after the agent registered against it.
Knowing what changed is not the same as knowing what breaks
Most teams run a schema diff on their API PR. That tells you a field was removed. It does not tell you which registered agents use that field, whether they call it as required, or what your blast radius is before the merge.
Knowing what changed is not the same as knowing what breaks.
This is the gap registry-level validation closes. It takes the breaking change signal from your PR and runs it against the registered consumers — your agent tool manifests, your MCP server schemas, your internal catalog entries — to surface actual impact before deployment, not after.
What registry-level validation looks like in practice
The input is two things: the proposed schema change (your base and head OpenAPI spec) and the registry entries you want to validate against — agent tool manifests, MCP tool schemas, or catalog entries.
The output is an assessment per registered consumer: does this change break anything they depend on, and if so, how severely.
The architecture is straightforward:
Agent tool manifest
↓
Tool discovery
↓
API registry
↓
CodeRifts registry validation
↓
API
For example: if your Payments API removes the transaction_id field from a response and your agent manifest declares that field as required output, the validation catches that as a CRITICAL break before the PR merges. If it adds a new required request field your agent does not send, same result.
The registry validates not just whether the schema is internally consistent, but whether it is safe for the consumers that depend on it.
The workflow that works
Open a PR that touches an OpenAPI spec. CodeRifts runs the standard breaking change analysis. If the change has agent impact, it runs registry validation automatically against any registered tool manifests in the repo. The PR comment surfaces both: what changed, who breaks, and a severity score.
If the registry validation returns CRITICAL, the PR is blocked by policy. The developer sees the exact field that breaks the exact agent that uses it, before anything ships.
This is what governance at the PR means in the context of agent infrastructure. Not a post-deploy alert. Not a runtime error. A pre-merge block with a specific reason.
Why this matters more now than six months ago
Agent adoption is accelerating faster than agent reliability tooling. Teams are registering tools against APIs without any mechanism to detect when those APIs change in breaking ways. The tools keep working — until they do not, and by then the failure is deep in a workflow that is hard to trace.
Observability tools tell you an agent failed. Registry-level validation prevents the failure from reaching an agent at all.
This is the layer missing in the agent stack.
Not runtime monitoring. Not API documentation. Not a diff report.
Contract guardrails at the registry layer — enforced before the merge.
Start catching agent-breaking API changes today
Install the GitHub App — free during beta, no CI config required.