Claude Mythos achieves 73% success rate on expert-level hacking challenges, completes full network takeover in 3 of 10 a
The UK's AI Safety Institute reports Claude Mythos Preview achieved a 73% success rate on expert-level capture-the-flag cybersecurity challenges and became the first AI model to complete a full 32-step simulated corporate network takeover, succeeding in 3 out of 10 attempts. The testing occurred in environments without active security monitoring or defenders.
Claude Mythos achieves 73% success rate on expert-level hacking challenges, completes full network takeover in 3 of 10 attempts
The UK's AI Safety Institute (AISI) reports that Anthropic's Claude Mythos Preview achieved a 73% success rate on expert-level cybersecurity challenges and became the first AI model to autonomously complete a full simulated corporate network attack from start to finish.
Benchmark performance
In capture-the-flag (CTF) evaluations with a 50 million token compute budget, Mythos Preview scored:
- 85% on apprentice-level tasks
- 95% on beginner-level tasks
- 93% on practitioner-level tasks
- 73% on expert-level challenges
According to AISI, no AI model could solve expert-level CTF tasks before April 2025. The institute places Mythos Preview in the top tier alongside GPT-5.4, Codex 5.3, and Claude Opus 4.6 for beginner-level performance.
Full network takeover simulation
AISI developed "The Last Ones" (TLO), a 32-step attack simulation against a simulated corporate network that would take human experts an estimated 20 hours to complete. With a 100 million token budget, Mythos Preview:
- Completed all 32 steps in 3 out of 10 attempts
- Averaged 22 of 32 steps across all attempts
- Outperformed Claude Opus 4.6, which averaged 16 steps
The model is the first to complete this end-to-end attack simulation, according to AISI.
Significant limitations in testing
AISI notes critical caveats: the test environments contained no active defenders, no security monitoring tools, and no consequences for actions that would trigger alarms on real networks. "There's no way to tell whether Mythos Preview could successfully breach a well-defended system" based on these results alone, the institute states.
The model also failed to complete a separate AISI simulation targeting industrial control systems, stalling in the IT network during earlier stages.
AISI concludes the model can "autonomously attack small, weakly defended and vulnerable enterprise systems where access to a network has been gained." Future evaluations will include hardened environments with active monitoring and incident response.
Limited availability and controversy
Anthropic launched Claude Mythos in early April but currently limits access to approximately 50 companies, reportedly due to cybersecurity concerns. Critics compare this to OpenAI's 2019 decision to restrict GPT-2, arguing the performance gains don't justify such limited access. Some speculate the restrictions are primarily for marketing purposes or due to compute capacity constraints.
What this means
This represents the first documented case of an AI model completing a full multi-stage network attack simulation autonomously. The 73% expert-level CTF score and ability to chain 22+ attack steps shows measurable advancement in AI cyber capabilities over 2025 models.
However, the absence of active defenses in testing leaves the practical threat level unclear. Real enterprise networks deploy endpoint detection, security monitoring, and incident response—none of which were present in AISI's simulations. The results highlight that AI models can now exploit basic security weaknesses at scale, reinforcing the importance of fundamental security practices like regular patching and strong access controls.
AISI and the UK's National Cyber Security Centre note these capabilities are dual-use: the same techniques that enable offensive operations could strengthen defensive cybersecurity systems.
Related Articles
UK AI Safety Institute confirms Claude Mythos finds more exploits as token spend increases
The UK's AI Safety Institute published an independent evaluation confirming Anthropic's Claude Mythos is highly effective at finding security vulnerabilities. The evaluation revealed a linear relationship: more tokens spent equals more exploits discovered, transforming security into an economic arms race.
Anthropic briefed Trump administration on Mythos model despite Pentagon lawsuit
Anthropic co-founder Jack Clark confirmed the company briefed the Trump administration on its Mythos model, which the company says is too dangerous for public release due to powerful cybersecurity capabilities. The briefing occurred despite Anthropic's ongoing lawsuit against the Department of Defense over AI system access restrictions.
Anthropic withholds Claude Mythos Preview from public release due to autonomous cybersecurity exploit capabilities
Anthropic has declined to publicly release Claude Mythos Preview, its most capable AI model, citing critical cybersecurity risks. Instead, the company launched Project Glasswing, providing controlled access to 50+ organizations including AWS, Apple, Google, and Microsoft, along with $100 million in usage credits and $4 million in direct donations to open-source security initiatives.
Anthropic restricts Claude Mythos access, exposing Europe's lack of AI safety infrastructure
Anthropic is restricting access to Claude Mythos Preview, a model the company claims can find security vulnerabilities better than most humans, to 52 technology partners. While the UK's AI Security Institute has already tested Mythos and published findings, most European cybersecurity agencies have limited or no access, revealing a structural gap in Europe's AI safety evaluation capacity.
Comments
Loading...