TakeOverBench

AI is rapidly getting more dangerous. We track progress towards an AI takeover scenario.

Dangerous capabilities timeline

According to "Model evaluation for extreme risks" (Shevlane, 2023), the most dangerous AI capabilities include Cyber-offense, Persuasion & manipulation, Political strategy, Weapons acquisition, Long-horizon planning, AI development, Situational awareness, and Self-proliferation. Progress on these capabilities is shown below.

Chart loading…

How do dangerous capabilities lead to a takeover?

Based on the literature, these are four plausible AI takeover scenarios.

We aim to use the best available benchmarks and the most plausible AI takeover scenarios, but the field of AI safety is rapidly developing. Some benchmark scores may therefore not perfectly reflect the actual associated dangerous capability, and some experts may disagree on the likeliness of specific threat models.