analysisOpenAI

OpenAI restricts cybersecurity AI access following Anthropic's model controls

TL;DR

OpenAI is restricting access to a new AI model with advanced cybersecurity capabilities to a small group of companies, mirroring Anthropic's decision to limit distribution of its Mythos Preview model. OpenAI's move builds on its February launch of the Trusted Access for Cyber pilot program following GPT-5.3-Codex, offering $10 million in API credits to participants.

2 min read
0

OpenAI Restricts Cybersecurity AI Access Following Anthropic's Model Controls

OpenAI is limiting access to a new AI model with advanced cybersecurity capabilities to a select group of companies, according to Axios reporting. The decision directly mirrors Anthropic's approach announced this week when it restricted access to Mythos Preview to only tech and security firms.

Current Programs and Models

OpenAI launched its "Trusted Access for Cyber" pilot program in February following the release of GPT-5.3-Codex, described as the company's most capable cybersecurity model to date. The pilot provides participants with access to particularly powerful models for defensive security work, backed by $10 million in API credits.

Anthropically's Mythos Preview model represents the company's latest cybersecurity-focused offering. The company has explicitly ruled out a public release, stating that models in the Mythos class will not ship to the broader market until adequate safety guardrails are in place.

Industry-Wide Safety Posture

Both companies cite the powerful hacking capabilities embedded in these models as the primary reason for restricting distribution. Rather than making advanced cybersecurity capabilities universally available, the companies are implementing staged rollouts to trusted partners in security and technology sectors.

Anthropic's decision to permanently exclude public access represents a stricter stance than OpenAI's approach, which does not rule out eventual broader availability. The timeline and conditions under which OpenAI might expand access to its new cybersecurity model remain undisclosed.

What This Means

The convergence of OpenAI and Anthropic on restricted access models signals growing industry consensus that certain AI capabilities pose sufficient security risks to warrant controlled distribution. This approach differs from each company's general model release strategy: while both companies maintain open or semi-open release policies for general-purpose models, specialized cybersecurity capabilities are receiving more cautious handling. The $10 million in credits for OpenAI's pilot suggests these capabilities remain valuable for legitimate defensive security work, but the restriction indicates both companies assess the offensive application risks as substantial enough to limit access during development and safety testing phases.

Related Articles

model release

Anthropic withholds Claude Mythos after finding thousands of OS vulnerabilities

Anthropic has announced Project Glasswing, restricting its new frontier model Claude Mythos Preview to defensive cybersecurity purposes through a coalition of 11 partners including AWS, Apple, Google, and Microsoft. The model has autonomously discovered thousands of high-severity vulnerabilities in major operating systems and web browsers—including a 27-year-old bug in OpenBSD and a 16-year-old vulnerability in FFmpeg—and can exploit them with 83.1% reliability on known vulnerabilities.

product update

OpenAI embeds Codex plugin directly into Anthropic's Claude Code

OpenAI released a plugin that embeds its Codex coding assistant directly into Anthropic's Claude Code, the market-dominant code IDE. The plugin offers standard code review, adversarial review, and background task handoff capabilities, requiring only a ChatGPT subscription or OpenAI API key.

product update

ChatGPT launches first streaming video app with Tubi for content discovery

OpenAI's ChatGPT has launched its first streaming video service integration, partnering with Tubi. The native app lets users search Tubi's catalog of over 300,000 movies and TV episodes using natural language queries.

analysis

AMD AI director reports Claude Code performance degradation since March update

Stella Laurenzo, director of AI at AMD, filed a GitHub issue documenting significant performance degradation in Claude Code since early March, specifically following the deployment of thinking content redaction in version 2.1.69. Analysis of 6,852 sessions with 234,760 tool calls shows stop-hook violations increased from zero to 10 per day, while code-reading behavior dropped from 6.6 reads to 2 reads per session.

Comments

Loading...