Why TakeOverBench exists
As artificial intelligence systems rapidly advance, we need clear, data-driven insights into their capabilities and the risks they may pose to human autonomy and control.
Our approach
The Theoretical Framework
To identify the most critical risks, we grounded our methodology in the research paper "Model evaluation for extreme risks" (Shevlane et al., 2023). This framework describes how advanced AI systems might develop 9 dangerous capabilities.
1. Mapping Capabilities
We evaluated dozens of public benchmarks and mapped the most rigorous and relevant ones to the 9 dangerous capabilities identified in Shevlane et al. (2023). For each benchmark, we describe why we selected it and what its limitations are.
2. Linking Scenarios
Capabilities don't exist in a vacuum. We composed various concrete Takeover Scenarios (threat models) and linked them to the specific capabilities required to execute them.
3. Evaluating Model Freshness
Due to the speed of AI development, many leading benchmarks become outdated within months. We track and highlight the "Months since latest update" to emphasize the critical need for regular benchmark runs on newer frontier models.
Take action
Understanding AI risks is the first step. Here's how you can contribute to ensuring AI remains beneficial and under human control:
Who we are
TakeOverBench is an initiative of two organizations dedicated to reducing existential risks from artificial intelligence and ensuring that advanced AI development remains under meaningful human control.

