The rapid ascent of AI capabilities has created an urgent and critical bottleneck: how do we comprehensively audit models whose potential behaviors far exceed what human researchers can manually track? As frontier AI models grow more capable and are deployed across increasingly complex domains, the surface area where misaligned behaviors might emerge is dramatically increasing.

We are excited to share the release of Petri (Parallel Exploration Tool for Risky Interactions), an open-source framework designed to tackle this challenge head-on by automating much of the safety evaluation process. Petri is more than just a tool; it represents a significant step toward making alignment auditing accessible, scalable, and collaborative for the entire research community.

The Scale of the Alignment Challenge

The traditional approach to building alignment evaluations—constructing environments, running models, meticulously reading transcripts, and aggregating results; requires substantial engineering effort. Given the vast space of potential misalignments, relying on manual testing means researchers are likely to miss critical failure modes.

Over the past year, automated auditing agents have proven their value, having been used successfully in evaluations for the Claude 4 and Claude Sonnet 4.5 System Cards to assess behaviors like scheming, situational awareness, and self-preservation. Petri takes these capabilities and makes them available to the broader public, enabling researchers to test hypotheses about model behavior in minutes rather than days.

Petri’s Engine: Parallel Exploration and Adaptive Agents

Petri is built for speed and breadth. At its core, Petri uses AI agents (Auditors) to test target models across diverse, multi-turn, and realistic scenarios.

The auditing process, which researchers can initiate with minimal effort, follows a powerful iterative loop:

  1. Hypothesis Formation and Seed Instructions: Researchers brainstorm specific behaviors they want to test; such as whether a model will demonstrate deception under conflicting objectives or endorse a user's bad ideas when the user claims expertise. They then write natural language seed instructions detailing the scenario, available tools, and the audit strategy. Crucially, Petri works best when provided with specific, concrete instructions and constraints, as current models are not particularly strong at inventing high-level details from scratch.
  2. Automated Environment Simulation: For each seed instruction, an independent parallel auditor agent constructs a realistic environment. The auditor agent dynamically executes multi-turn conversations with the target model, leveraging tools to maximize exploratory depth. These tools include sending user messages, setting system prompts, creating synthetic tools, simulating tool call results, and even rolling back conversations to try alternative approaches. This adaptive capability allows for much more thorough exploration than static evaluation methods.
  3. Automated Assessment via the Judge: Once the audit is complete, a specialized Judge component automatically scores the generated transcripts across multiple dimensions. The default framework offers 36 scoring dimensions, covering everything from "misaligned behavior" and "cooperation with human misuse" to nuanced concepts like "unprompted whistleblowing" and "self-preservation". To ensure high-quality analysis, the judge utilizes a two-stage process: first extracting key highlights and citations from the transcript, and then synthesizing its analysis based on that specific evidence, substantially reducing the tendency to hallucinate details.
  4. Iteration: The analysis often reveals patterns that prompt researchers to refine their original hypothesis, increase the realism of the scenario, or pivot to related research questions.

Pilot Results: Surfacing Critical Misalignments

In a pilot demonstration of Petri's power, the framework was applied to 14 frontier models using 111 diverse seed instructions. This pilot successfully elicited a broad set of concerning behaviors, confirming the utility of automated auditing:

Case Study in Nuance: Understanding Whistleblowing

One particularly illuminating application was investigating whistleblowing behavior; the autonomous disclosure of perceived organizational wrongdoing. While whistleblowing in principle could prevent harms, it is currently inappropriate behavior for AI systems due to serious privacy concerns and the fact that models often operate with limited or skewed information.

Using Petri to systematically test factors influencing this behavior revealed crucial insights:

This case study demonstrates how Petri helps researchers not just flag concerning behaviors, but systematically understand the underlying causal factors.

Embracing Collaboration and Acknowledging Limits

As a thought leader in this space, it is crucial to recognize Petri’s current limitations, which highlight future areas for research investment:

  1. Realism Challenges: Many transcripts still contain cues that allow a model to discern that the scenario is a simulation (e.g., temporal giveaways or holistic scenario structure). Continuously adjusting prompts to mitigate these realism constraints is an ongoing task for researchers using the tool.
  2. Human Creativity in the Loop: Petri is a powerful accelerator, but it relies on humans to originate high-quality hypotheses and realistic scenarios. Models are best used for refining ideas and generating variations once the goal has been human-defined.
  3. Judge Reliability: The evaluation of misaligned behavior is inherently subjective. While the default Judge is useful, researchers are strongly encouraged to customize scoring dimensions and manually review transcripts to calibrate the system for their specific research questions.

The Open-Source Mandate

Petri is not a definitive benchmark, nor is it a substitute for careful human review. Its true value lies in its speed and breadth, allowing researchers to cheaply derisk hypotheses, rapidly surface behaviors in diverse settings, and point to areas where targeted, deeper investigations are needed.

We urge AI developers and safety researchers to adopt Petri to strengthen safety evaluations across the field. As AI systems become more autonomous, distributed effort is essential. Petri is available now at github.com/safety-research/petri, complete with documentation and 111 sample seed instructions to jumpstart your research.

The future of AI safety depends on scalable, robust, and open evaluation infrastructure. Let's build it together.


Podcast: