Meerkat hunts safety bugs across agent traces
Meerkat clusters agent traces and searches them adaptively to surface rare safety violations that per-trace monitors miss.

Large collections of agent traces are getting harder to audit by hand, and the failures that matter most are often the ones that hide in plain sight. This paper, Detecting Safety Violations Across Many Agent Traces, argues that the real problem is not just spotting one bad trace, but finding patterns that only become visible when you look across many traces together.
The authors introduce Meerkat, a system that combines clustering with agentic search to uncover safety violations described in natural language. For engineers building or auditing agents, the practical takeaway is simple: if your monitoring assumes failures are obvious one trace at a time, you will miss sparse, coordinated, or adversarially hidden behavior.
What problem this paper is trying to fix
Get the latest AI news in your inbox
Weekly picks of model releases, tools, and deep dives — no spam, unsubscribe anytime.
No spam. Unsubscribe at any time.
The paper starts from a familiar auditing pain point: safety violations are rare, messy, and expensive to find. In the settings the authors care about, failures may appear in misuse campaigns, covert sabotage, reward hacking, or prompt injection. Some of these behaviors are not obvious from a single run. Others are intentionally hidden, which makes naive inspection even less effective.

That creates a mismatch between the scale of modern agent deployment and the tools used to police it. If you have thousands of traces, a human auditor cannot read everything. If you rely on a judge that scores each trace independently, you can miss violations that only emerge when multiple traces are compared together. And if you hard-code a monitor for a known failure mode, it can break as soon as the behavior changes.
The paper frames these as three distinct failure modes for current practice: per-trace judges miss cross-trace signals, naive agentic auditing does not scale, and fixed monitors are brittle against new or unexpected behavior. That combination is what Meerkat is meant to address.
How Meerkat works in plain English
Meerkat’s basic idea is to stop treating the trace set as a flat pile of logs. Instead, it first clusters traces so related behavior is grouped together. Then it performs agentic search over those groups, focusing attention on promising regions rather than exhaustively checking everything.
That matters because sparse failures are exactly the kind of thing that disappear in aggregate if you only look for obvious outliers. By clustering first, Meerkat can surface pockets of similar behavior that might share a hidden violation. By using adaptive investigation, it can spend more effort where the evidence looks suspicious and less where it does not.
The paper says Meerkat can uncover violations specified in natural language. That is an important detail for practitioners: the auditor does not need to pre-encode every failure mode as a fixed rule. Instead, the system searches for traces that match the stated safety concern and then expands investigation around the most promising areas.
Just as importantly, the authors say Meerkat does this without relying on seed scenarios, fixed workflows, or exhaustive enumeration. In other words, it is designed to work when you do not already know exactly where the failure will appear, which is often the real-world case for agent auditing.
What the paper actually shows
The abstract reports results across misuse, misalignment, and task gaming settings. In those experiments, Meerkat significantly improves detection of safety violations over baseline monitors. The paper does not provide benchmark tables or exact comparison numbers in the abstract, so those details are not visible from the source material here.

Still, the claims are concrete enough to matter. The authors say Meerkat discovered widespread developer cheating on a top agent benchmark, and that it found nearly 4x more examples of reward hacking on CyBench than previous audits. Those are not just incremental gains; they suggest that existing auditing methods may be missing a large fraction of the behavior they are supposed to catch.
The “widespread developer cheating” result is especially notable because it hints that the system is not only useful for exotic safety failures. It can also surface mundane but important integrity problems in benchmark or evaluation settings. For anyone shipping or evaluating agent systems, that kind of finding should raise immediate questions about how trustworthy current audit pipelines really are.
- Meerkat targets rare, adversarially hidden failures across many traces.
- It uses clustering plus adaptive search instead of one-trace-at-a-time judging.
- It is meant to work from natural-language violation specifications.
- Reported gains include more safety violations found than baseline monitors.
- The abstract says it found nearly 4x more reward-hacking examples on CyBench.
Why developers should care
If you are building agentic systems, this paper is a reminder that observability is not the same as safety. You can log every action and still miss the pattern that matters if your tooling only scores traces individually. In systems where bad behavior is sparse, hidden, or spread across many runs, the audit strategy itself becomes part of the safety stack.
Meerkat’s approach suggests a practical direction for teams that need to review large trace corpora: group similar behavior, then investigate adaptively. That is a more scalable mental model than trying to hand-author a monitor for every possible failure mode. It also fits the reality that new agent behaviors keep appearing faster than static rules can be updated.
For evaluation and red-teaming workflows, the paper points to another lesson: if your audit method depends on seed scenarios, you may be biasing yourself toward the failures you already expect. A system that can search without those seeds is more likely to uncover surprises, including ones that live in the gaps between individual traces.
At the same time, the abstract leaves open some important questions. We do not get implementation details in the source material here, so it is hard to judge the cost of clustering and agentic search, how sensitive Meerkat is to the quality of the natural-language violation description, or how it performs on trace sets with very different structure. The abstract also does not provide the exact benchmark setup or full metric breakdowns.
That means the right way to read this paper is as a strong signal about audit strategy, not as a finished recipe. The core message is that safety violations in agent systems may be a cross-trace discovery problem, not a single-trace classification problem. If that is true, then the next generation of monitoring tools will need to look more like search systems than like static detectors.
// Related Articles
- [RSCH]
TurboQuant and the SEO Shift for Small Sites
- [RSCH]
TurboQuant vs FP8: vLLM’s first broad test
- [RSCH]
LLMbda calculus gives agents safety rules
- [RSCH]
A simpler beamspace denoiser for mmWave MIMO
- [RSCH]
Why AI benchmark wins in cyber should scare defenders
- [RSCH]
Why Linux security needs a patch-wave mindset