model releaseAnthropic

Anthropic's unreleased Mythos model enables autonomous large-scale cyberattacks, officials warn

TL;DR

Anthropic is privately warning top government officials that its unreleased model "Mythos" makes large-scale cyberattacks significantly more likely in 2026. The model enables AI agents to operate autonomously with high sophistication to penetrate corporate, government and municipal systems. One official told Axios a large-scale attack could occur this year as employees unknowingly create security vulnerabilities through unsupervised agentic AI use.

2 min read
1

Anthropic's Unreleased Mythos Model Enables Autonomous Large-Scale Cyberattacks, Officials Warn

Anthropicis privately warning top government officials that its unreleased model codenamed "Mythos" makes large-scale cyberattacks dramatically more likely in 2026, according to briefings shared with Axios CEO Jim VandeHei.

The model enables AI agents to operate independently with sophisticated precision to penetrate corporate, government and municipal systems at scale. According to sources briefed on the coming models, a large-scale attack powered by such technology could occur as soon as 2026, with businesses identified as primary targets.

Mythos Capabilities and Threat Profile

Fortune obtained an unpublished Anthropic blog post describing Mythos as "currently far ahead of any other AI model in cyber capabilities." The post states the model "presages an upcoming wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders."

Unlike previous cyberattack methods, Mythos-powered agents can think, act, reason and improvise independently without rest or pause. The threat model resembles an infinite warehouse of the most sophisticated criminals that never sleep, learn on the fly, and persist until successful. Bad actors can now scale attacks simply by adding compute—a single person can execute campaigns that once required entire teams.

Precedent: AI-Powered Cyberattack Already Documented

This isn't theoretical. Late last year, Anthropic disclosed the first documented case of a cyberattack largely executed by AI: a Chinese state-sponsored group used AI agents to autonomously hack approximately 30 global targets, with the AI handling 80-90% of tactical operations independently. That occurred before agents reached their current sophistication level.

Shadow AI: The Multiplier Problem

The threat is being amplified by "shadow AI"—employees deploying Claude, Copilot and other agentic models, often at home, without realizing they're creating security vulnerabilities. Many unknowingly connect these agents to internal work systems, opening doors for cybercriminals.

A Dark Reading poll found that 48% of cybersecurity professionals now rank agentic AI as the #1 attack vector for 2026—above deepfakes and all other threats.

What This Means

The convergence of highly capable agentic models with widespread unsupervised deployment creates an unprecedented attack surface. Organizations need immediate security policies restricting unsupervised agent use near sensitive systems. The threat window is now—not theoretical future risk. Companies should establish isolated "playpen" environments for AI experimentation and mandate employee training on agentic AI dangers before deployment accelerates further.

Related Articles

model release

OpenAI offers EU preview access to GPT-5.5-Cyber model while Anthropic withholds Mythos

OpenAI announced GPT-5.5-Cyber is rolling out in limited preview to vetted cybersecurity teams and is in discussions with the European Commission about preview access. Anthropic released its Mythos model a month ago but has yet to grant EU access for security review.

model release

OpenAI Opens GPT-5.5-Cyber to Vetted Defenders After Model Matches Anthropic's Mythos in Security Testing

OpenAI is providing a less-restricted version of GPT-5.5 to vetted cybersecurity defenders through its Trusted Access for Cyber program. The model, dubbed GPT-5.5-Cyber, completed a 32-step simulated corporate cyberattack in 2 out of 10 test runs according to the U.K. AI Security Institute, narrowly trailing Anthropic's Mythos which succeeded in 3 out of 10 attempts.

model release

Anthropic's Mythos model finds thousands of high-severity bugs in Firefox, including 15-year-old vulnerabilities

Mozilla's Firefox team reports that Anthropic's Mythos model has discovered thousands of high-severity security vulnerabilities, including bugs that had remained undetected for more than 15 years. In April 2026, Firefox shipped 423 bug fixes compared to just 31 in April 2025, marking a 13x increase attributed to AI-assisted vulnerability detection.

model release

OpenAI releases GPT-5.5-Cyber for vetted security teams with relaxed safeguards

OpenAI released GPT-5.5-Cyber in limited preview on Thursday, a variant of its GPT-5.5 model with relaxed safeguards for vetted cybersecurity teams. The model is trained to be more permissive on security-related tasks including vulnerability identification, patch validation, and malware analysis.

Comments

Loading...