Skip to content
FAQ

Anthropic's 'Claude Mythos' Leaked: The Most Powerful AI Model Yet — and a Cybersecurity Threat

A CMS misconfiguration exposed Anthropic's secret model codenamed 'Mythos', positioned above Opus in a new 'Capybara' tier. The company is now privately briefing US officials about its unprecedented cybersecurity risks while conducting limited early access testing.

5 min read

A configuration error at Anthropic’s content management system inadvertently revealed something the company had been keeping under wraps: a new AI model of unprecedented capability, codenamed “Mythos.” The accidental exposure — which surfaced roughly 3,000 unpublished internal assets on March 26 — has triggered a rare sequence of events that shows just how seriously the AI frontier is taking the dual-edged nature of its most powerful systems.

The Leak That Wasn’t Supposed to Happen

The CMS misconfiguration exposed internal documents, product roadmap materials, and benchmark data that Anthropic had not yet made public. Among those materials was evidence of a model operating in a product tier codenamed “Capybara” — positioned above the company’s current flagship Claude Opus 4.6 in the product hierarchy.

Anthropic moved quickly to contain the exposure, but security researchers and journalists had already captured enough to piece together a picture of what Mythos is — and what it’s capable of. Within days, Anthropic confirmed the model’s existence to Fortune, describing it as “by far the most powerful AI model we have ever developed.”

That’s a notable statement from a company whose Opus 4.6 already sits at or near the top of most independent AI benchmarks. But Mythos, according to internal documentation, doesn’t just incrementally advance the state of the art — it represents what Anthropic internally calls “a step change in capabilities.”

Dramatically Superior Benchmarks

The leaked data shows Mythos dramatically outperforming Claude Opus 4.6 across three core domains: coding, academic reasoning, and cybersecurity. On SWE-bench Verified — the gold standard for autonomous software engineering capability — Mythos reportedly scores significantly above any currently released model. On advanced academic benchmarks like GPQA Diamond and AIME 2025, the performance gaps are described as “substantial” in internal materials.

But it’s the cybersecurity benchmarks that have attracted the most attention, and the most concern.

Mythos is described as capable of identifying and exploiting vulnerabilities at a scale that “far outpaces the efforts of defenders.” In internal red-teaming exercises, the model demonstrated the ability to autonomously chain multiple steps of a cyberattack — reconnaissance, vulnerability identification, exploitation, and lateral movement — without requiring human prompting at each stage.

This reflects Mythos’s core design philosophy: it is an agentic model by default, built to plan and execute multi-step task sequences end-to-end. Where Claude Opus 4.6 benefits from agentic scaffolding but was fundamentally designed as a conversational assistant, Mythos is architected from the ground up to operate autonomously across extended task horizons.

Briefing Washington Before Launch

What makes the Mythos situation unusual — even by frontier AI standards — is the step Anthropic took before announcing general availability or even a release date: private briefings with senior US government officials.

According to reporting from Axios, Anthropic has been quietly briefing officials in the intelligence community and relevant congressional committees about Mythos’s potential for offensive cybersecurity applications. The company is framing this proactively, positioning itself as a responsible actor flagging dual-use risks before the model reaches broad deployment.

It’s a playbook that reflects Anthropic’s safety-first brand positioning, but also the genuine unease that senior staff reportedly feel about releasing a model with these capabilities into a world where adversarial actors — state-sponsored and otherwise — are actively looking for AI-powered offensive tools.

The briefings also carry a practical dimension: Anthropic is reportedly interested in positioning Mythos for classified government work, particularly in cyber defense applications. The company already has existing relationships with DARPA and the intelligence community through its broader research partnerships.

Cybersecurity Partners Get First Access

Rather than a broad public launch, Anthropic is currently running Mythos through a limited early access program specifically scoped to cybersecurity partners. This group includes a small number of vetted organizations working on defensive AI applications — intrusion detection, vulnerability assessment, threat intelligence — where Mythos’s capabilities can be applied in controlled, monitored environments.

This approach mirrors how OpenAI handled its own most sensitive model releases, creating tiered access structures that allow the company to gather real-world capability data while limiting exposure of the most dangerous capabilities to vetted actors.

Notably, no general release date has been announced. This is itself unusual: Anthropic typically pre-announces model releases with some lead time to prepare the developer ecosystem. The absence of a public timeline for Mythos suggests either that safety testing is still ongoing, or that the company is navigating a more complex policy and legal landscape than it has for previous releases.

A New Pricing Tier and the Economics of Frontier AI

When Mythos does launch publicly, it will introduce a fourth product tier above Opus — what the leaked materials call the “Capybara” tier. This is likely to be Anthropic’s most expensive API offering by a significant margin, given the compute requirements and market positioning of a model this capable.

The business logic is clear: as the gap between frontier AI and the open-source alternatives continues to widen on the most demanding tasks, the willingness of enterprise customers to pay premium prices for genuine capability advantages grows as well. Mythos appears designed to be the offering that justifies a price point that no current AI product can claim.

This also has implications for Anthropic’s competitive posture. The company has been in an uncomfortable position in early 2026 — Claude Opus 4.6 remains highly regarded, but Google’s Gemini 3.1 Pro has been leading on many benchmarks, and OpenAI’s GPT-5.x series has successfully completed its transition away from GPT-4o. Mythos, if it performs as internal data suggests, would re-establish Anthropic at the technical frontier in a way that changes the competitive calculus for the entire industry.

The Dual-Use Problem at Unprecedented Scale

The deeper issue that Mythos raises — one that industry observers are beginning to grapple with — is what it means for the safety and policy ecosystem when a model is powerful enough that its offense-defense balance tips meaningfully toward offense.

Previous AI capabilities — coding assistants, image generation, even earlier agentic systems — presented manageable dual-use concerns. Regulatory and industry frameworks developed quickly enough to keep pace. Mythos, as described in the leaked materials, is different in degree to a point where it might be different in kind.

An AI system that can autonomously chain cyberattack steps at scale doesn’t just make individual bad actors more effective — it potentially changes the economics of state-sponsored cyber operations, lowers the bar for sophisticated attacks to a level where much smaller actors can mount them, and creates novel attack surfaces that existing detection systems weren’t designed to identify.

Anthropic’s decision to brief government officials before launch is recognition of this reality. The question the broader industry is now asking is whether that recognition is enough — or whether Mythos represents the moment when AI capabilities require a fundamentally different governance framework than we currently have.

For now, Anthropic is proceeding carefully. But the existence of Mythos — however accidentally revealed — means that a new threshold in AI capability has already been crossed, whether or not the model is yet in anyone’s hands.

Anthropic Claude Mythos AI safety cybersecurity foundation models AI agents
Share

Related Stories

The End of GPT-4o: OpenAI Completes Model Retirement as GPT-5.4 Becomes the New Standard

OpenAI completed the full retirement of GPT-4o on April 3, 2026, alongside GPT-4.1 and o4-mini. Only 0.1% of daily users were still choosing GPT-4o at the time of retirement. GPT-5.4 — available in Standard, Thinking, and Pro variants — is now the platform baseline, while Gemini 3.1 Pro leads 13 of 16 major benchmarks at roughly one-third of the API cost.

5 min read