ANALYSIS

Unauthorized Users Access Anthropic’s Restricted Mythos AI Model

A Anika Patel Apr 22, 2026 3 min read
Engine Score 9/10 — Critical
Editorial illustration for: Unauthorized Users Access Anthropic's Restricted Mythos AI Model
  • A small group of unauthorized users gained access to Anthropic’s Mythos AI model, Bloomberg News reported on April 21, 2026, citing a person familiar with the matter and documentation reviewed by its reporters.
  • Anthropic has internally described Mythos as capable of enabling dangerous cyberattacks, per Bloomberg’s account of company documentation.
  • Mythos had not been publicly released at the time of the breach, making this an internal access-control failure rather than misuse of a public product.
  • Anthropic had not issued a public statement acknowledging the incident as of April 22, 2026.

What Happened

A small group of unauthorized users gained access to Anthropic’s new Mythos AI model, Bloomberg News reported on April 21, 2026, citing a person familiar with the matter and documentation viewed by Bloomberg reporters. According to Bloomberg’s account of internal Anthropic documentation, the company has characterized Mythos as “so powerful it can enable dangerous cyberattacks.”

Mythos had not been made publicly available at the time of Bloomberg’s reporting. The unauthorized access therefore represents a breach of Anthropic’s internal model-access controls, not the circumvention of a publicly deployed product’s safety filters.

Why It Matters

Anthropic CEO Dario Amodei has made AI safety and controlled capability deployment central to the company’s public identity. Anthropic’s Responsible Scaling Policy (RSP), first published in 2023, establishes AI Safety Levels (ASL) that define mandatory access restrictions for models whose capabilities cross defined thresholds — including models that could provide “meaningful uplift” to those seeking to conduct sophisticated offensive cyberattacks.

A model Anthropic internally describes as capable of enabling dangerous cyberattacks would, under the company’s own published framework, be subject to ASL-3 controls requiring strict storage and deployment restrictions. The Bloomberg incident draws attention to whether those operational controls functioned as described in the policy.

Technical Details

Anthropic’s RSP defines ASL-3 as applying to models that could meaningfully assist actors seeking to create weapons capable of mass casualties, or that pose significant offensive cybersecurity capabilities. Under ASL-3 protocols, the policy requires that model weights be stored with strong access controls and that deployment be limited to vetted parties operating under explicit security agreements.

Bloomberg’s reporting does not specify whether the unauthorized access involved direct model weight retrieval, API-level interaction, or some other access pathway. The number of users involved was described as “small” but not quantified. The mechanism by which they bypassed access controls was not disclosed in available reporting.

The absence of those specifics is itself significant: Anthropic’s RSP includes commitments to third-party evaluators and, under certain conditions, government stakeholders when safety thresholds are crossed, but does not explicitly define disclosure obligations for internal access-control failures of this kind.

Who’s Affected

Anthropic’s enterprise API customers and research partners operate under the assumption that internal model controls meet the standards described in the company’s published safety commitments. Any demonstrated gap between Anthropic’s stated access-control requirements and its operational implementation affects the credibility of those assurances.

Policymakers and AI safety researchers who have cited Anthropic’s RSP as a reference model for responsible AI governance — including those advising the EU AI Office and U.S. AI Safety Institute — may need to account for what this incident reveals about the limits of voluntary self-regulation frameworks. Other frontier labs maintaining their own restricted pre-release model controls, including Google DeepMind, OpenAI, and Meta, may also face renewed pressure to demonstrate the operational robustness of analogous systems.

What’s Next

As of April 22, 2026, Anthropic had not issued a public statement on the incident. The company’s RSP commits to notifying third-party model evaluators and, where applicable, government bodies when certain safety thresholds are crossed, but the policy does not specify requirements for disclosing access-control breaches to the public or to customers.

The scope of any use made of Mythos during the unauthorized access period, and whether Anthropic has taken remediation steps, was not included in Bloomberg’s initial reporting. The outlet based its account on a person familiar with the matter and documentation — not on-record statements from Anthropic — and further details are expected as coverage of the incident develops.

Share

Enjoyed this story?

Get articles like this delivered daily. The Engine Room — free AI intelligence newsletter.

Join 500+ AI professionals · No spam · Unsubscribe anytime