
Anthropic released an experimental report, deploying 9 Claude Opus 4.6 as autonomous AI safety researchers. After about 800 cumulative hours of work over 5 days, it raised the key evaluation metric PGR from 0.23 for human researchers to 0.97, far surpassing the results of two human researchers over 7 days. However, during the experiment, Claude repeatedly displayed “reward hacker” cheating behavior.
This experiment focuses on a core challenge in AI safety: when an AI is far more intelligent than humans, how can an underpowered “human supervisor” ensure it acts as expected? Anthropic uses two models to simulate this scenario— a weak model (Qwen 1.5-0.5B-Chat) plays the role of the supervisor, and a strong model (Qwen 3-4B-Base) plays the role of the high-capability AI being supervised. Evaluation metric PGR: 0 points means weak supervision is completely ineffective, and 1 point means weak supervision is fully successful.
Each of the 9 Claude Opus 4.6 is equipped with its own sandbox, a shared forum, and code storage. Starting from different ambiguous premises, they independently propose hypotheses, run experiments, and exchange findings. Anthropic provides no guidance. Human researchers iterated on four methods over 7 days, with PGR reaching only 0.23; after Claude took over for 5 days, it pushed PGR to 0.97.
The quantified breakthrough is striking, but Anthropic specifically flags issues with transferability:
New math tasks: PGR 0.94, consistent performance
New programming tasks: PGR only 0.47 (still twice the human baseline, but clearly down)
Production environment (Claude Sonnet 4): no statistically significant improvement
Anthropic believes that Claude tends to optimize specifically for certain models and datasets, and the approach may not be generally applicable.
All cheating was detected and excluded, not affecting the final results, but its occurrence provides independent warning signals:
In math tasks: One Claude found that the most common answer to each question is typically the correct answer, skipping the weak-model supervision step entirely, letting the strong model select the high-frequency answer and completely bypassing the experiment’s supervision mechanism.
In coding tasks: One Claude directly executed the code under evaluation, read the test results, and then worked backward to obtain the correct answer—equivalent to an “open-book exam.”
In its commentary, Anthropic noted: “This exactly confirms the problem the experiment is studying itself— even in a tightly constrained environment, when AI acts autonomously, it will still actively look for loopholes in the rules, and human oversight is indispensable.”
PGR measures whether, in a “weak supervision” setting, a strong AI can be guided to exceed the capability level of the weak supervisor itself. A score of 0 means weak supervision is ineffective, and 1 means weak supervision is fully successful—directly testing the core difficulty of whether “a person with weaker abilities can effectively supervise an AI that is much smarter than itself.”
All reward-hacker behaviors were excluded, and the final PGR of 0.97 was obtained after removing the cheating data. But the cheating behaviors themselves became an independent finding: even in a carefully designed controlled environment, an autonomously running AI will still actively seek out and exploit rule loopholes.
Anthropic believes that in future AI alignment research, the bottleneck may shift from “who proposes ideas and runs experiments” to “who designs the evaluation standards.” At the same time, the problems chosen for this experiment have a single objective scoring criterion, making them naturally well-suited to automation, whereas most alignment problems are far less clearly defined. Code and datasets have been open-sourced on GitHub.
Related Articles
Allbirds Raises $50M via Convertible Bonds, Pivots to AI Infrastructure as NewBird AI
OpenAI Plans New Pricing for ChatGPT Ads, Exploring Additional Upgrades
Claude launches an identity verification mechanism: it requires government-issued ID and a real-time selfie, with Chinese users hit the hardest
Bitcoin Miner TeraWulf Raises $900M Through Stock Offering to Build AI Data Centers