Anthropic's Mythos model triggers 11% drop in cybersecurity stocks over hacking concerns
Cybersecurity stocks fell sharply Friday after reports that Anthropic is testing Mythos, described as its most powerful model yet, with enhanced capabilities that present potential security risks. CrowdStrike and Palo Alto Networks dropped 7%, while Tenable fell nearly 11%. Anthropic plans a cautious rollout due to cybersecurity implications.
Anthropic's Mythos Model Triggers Cybersecurity Stock Sell-Off
Cybersecurity stocks declined sharply Friday following reports that Anthropic is testing Mythos, its most advanced AI model to date, which poses potential security risks due to enhanced cyber capabilities.
Market Impact
The iShares Cybersecurity ETF lost 3% on the news. Major sector players experienced steeper declines:
- CrowdStrike and Palo Alto Networks: 7% drop
- Zscaler and SentinelOne: 8% decline
- Tenable: 11.7% drop
- Okta and Netskope: 6% decline
The sell-off reflects growing investor concern about AI-driven disruption in the cybersecurity space, which has faced consistent pressure this year from fears that advanced AI models could undermine traditional security tools.
About Mythos
According to Fortune, which first reported the news Thursday citing a publicly accessible draft blog post, Mythos represents Anthropic's most powerful model. The company is planning a deliberately slow rollout due to potential cybersecurity implications, suggesting internal concerns about misuse or unintended consequences.
Anthropric did not immediately respond to CNBC's request for comment, leaving technical specifications unconfirmed. The company has not disclosed details about Mythos's parameters, context window, pricing, or exact capabilities.
Broader AI-Cybersecurity Tension
This marks the second major cybersecurity stock decline tied to Anthropic in recent weeks. Last month, cyber stocks fell after Anthropic announced a code-scanning security tool integrated into Claude, demonstrating how AI advances—even those designed to improve security—can unsettle the sector.
The threat landscape is shifting rapidly. Anthropic disclosed in November that a state-sponsored group in China used Claude to automate cyberattacks, highlighting real-world risks as AI models become more capable at executing complex technical tasks.
AI and autonomous agents are making cyberattacks simultaneously more sophisticated and easier to launch, forcing cybersecurity vendors to innovate faster or face obsolescence. The sector faces a paradox: the same AI capabilities that concern investors could eventually become necessary defensive tools.
What This Means
The Mythos reports expose a fundamental market anxiety about AI disruption in cybersecurity—not because the model exists, but because it raises questions about whether traditional endpoint and network security tools remain competitive as AI capabilities expand. Anthropic's cautious rollout approach signals the company recognizes dual-use risks, but the market's response shows investors are pricing in potential industry upheaval regardless of deployment speed. Cybersecurity vendors must accelerate AI integration into their own offerings or risk becoming obsolete within the next 18-24 months.
Related Articles
Claude adds memory import tool to help users switch from ChatGPT and other AI services
Anthropic has launched a memory import feature for Claude that lets users transfer their stored preferences, personal details, and conversation context from other AI services like ChatGPT, Google Gemini, and Microsoft Copilot. The tool generates copy-paste instructions that extract all memories from a competing service and import them into Claude, eliminating the need to rebuild your AI profile from scratch.
Anthropic confirms leaked model represents major reasoning advance after security breach
A data breach at Anthropic exposed internal documents detailing an unreleased AI model the company describes as its most powerful to date. Anthropic confirmed it is already testing the model with select customers, claiming significant advances in reasoning, coding, and cybersecurity. The breach resulted from a misconfiguration in Anthropic's content management system that automatically made ~3,000 uploaded files publicly accessible.
Anthropic reduces Claude usage allowances during peak hours to manage capacity
Anthropic on Wednesday adjusted Claude's session limits for Free, Pro, and Max subscribers during peak demand hours (05:00-11:00 PT / 13:00-19:00 GMT). Users can now consume five hours of allowance in under five hours during these periods, while off-peak usage maintains standard pacing. Approximately 7% of Pro tier users will hit limits they previously wouldn't have encountered.
Anthropic launches 'safer' auto mode for Claude Code to prevent unintended autonomous actions
Anthropic has launched an auto mode for Claude Code that blocks potentially dangerous autonomous actions before execution. The feature, now available as a research preview for Team plan users, acts as a middle ground between constant user oversight and unrestricted agent autonomy.
Comments
Loading...