Anthropic launches Claude Code 'auto mode' with AI-powered permission classifier
Anthropic has released 'auto mode' for Claude Code, a permissions system that sits between conservative defaults and fully disabled safeguards. The feature uses a classifier to automatically approve safe actions like file writes and bash commands while blocking potentially destructive operations.
Anthropic Launches Claude Code 'Auto Mode' With AI-Powered Permission Classifier
Anthropric rolled out "auto mode" for Claude Code on March 24, 2026, introducing a new permissions framework that balances developer convenience against safety risks.
The feature addresses a usability problem: Claude Code's default configuration requires explicit user approval before executing each file write or bash command. Developers seeking faster execution have historically disabled all permissions using the --dangerously-skip-permissions flag, creating significant security exposure.
Auto mode introduces a middle path using a machine learning classifier that pre-screens each tool invocation before execution. The classifier identifies potentially destructive actions—including mass file deletion, sensitive data exfiltration, and malicious code patterns—and blocks them automatically. Actions deemed safe proceed without user interruption. If Claude repeatedly attempts blocked actions, the system escalates to a user permission prompt.
Anthropric explicitly notes that auto mode reduces risk compared to fully disabled permissions but does not eliminate it entirely. The company recommends using auto mode exclusively in isolated development environments.
Rollout Timeline
Claude Teams users gained access to auto mode as a research preview on March 24. Enterprise and API customers will receive access within days, according to Anthropic's announcement.
This update follows Anthropic's unveiling of a separate research preview feature that enables Claude to control macOS directly—another capability gated behind safety controls.
What This Means
Auto mode addresses a genuine friction point in AI-assisted development: the trade-off between safety guardrails and operational efficiency. By delegating routine safety checks to an ML classifier, Anthropic reduces manual approval overhead while maintaining the ability to catch genuinely dangerous operations. However, the existence of a classifier that can be circumvented introduces new attack surface—adversarial prompts could potentially exploit classification boundaries. The elevation to manual permission prompts when Claude insists on blocked actions suggests the system relies on Claude's own behavior modification rather than hard technical barriers, which may be bypassable. Enterprise adoption will likely depend on how well the classifier generalizes to production codebases with domain-specific patterns.
Related Articles
Anthropic Doubles Claude Code Rate Limits, Secures 300+ MW Compute from SpaceX's Colossus 1
Anthropic has secured access to all compute capacity at SpaceX's Colossus 1 data center, adding more than 300 megawatts of new capacity within the month. As a result, the company is doubling five-hour rate limits for paid Claude Code users and removing peak hour restrictions for Pro and Max tiers.
Anthropic doubles Claude Code usage limits for paid users, increases API capacity by up to 1500%
Anthropic has doubled Claude Code's five-hour usage limits for Pro, Max, Team, and Enterprise users while removing peak hour restrictions for Pro and Max plans. The company also increased API limits by up to 1500% for input tokens per minute through a compute capacity deal with SpaceX's Colossus 1 data center.
OpenAI launches Trusted Contact feature allowing ChatGPT to alert designated friends during suicide risk
OpenAI has launched Trusted Contact for ChatGPT, allowing users 18+ to designate one adult contact who can be notified if the company's trained human review team detects serious self-harm risk. The feature comes after over 1 million of ChatGPT's 800 million weekly users expressed suicidal thoughts in conversations, and follows a 2025 wrongful death lawsuit.
GitHub Reduces Token Usage in Copilot Agentic Workflows Running on Pull Requests
GitHub has optimized token usage in its production agentic workflows that run on every pull request. The company instrumented its own Copilot workflows to identify inefficiencies and built agents to address them, aiming to reduce accumulated API costs.
Comments
Loading...