Anthropic announced its most capable model ever and refused to ship it. Claude Mythos Preview can autonomously find and exploit software vulnerabilities at a level Anthropic says rivals skilled human security researchers. It has already turned up thousands of high-severity zero-days in every major operating system and browser, including flaws that had survived 27 years in OpenBSD. Instead of a public release, Anthropic gave 52 vetted organizations gated access through Project Glasswing, committed $100M in usage credits, and shipped Claude Opus 4.7 as the public-facing model instead. The Fed and Treasury are now briefing US bank CEOs. The UK AI Safety Institute's own evaluation confirms the capability claims are real.
- Anthropic calls it 'the greatest alignment-related risk of any model we have released to date.'
- First AI model to clear UK AISI's expert-level capture-the-flag challenges at a 73% pass rate - no prior model had broken 50%.
- First AI model to complete UK AISI's 32-step enterprise-network attack simulation end-to-end.
- Found thousands of high-severity zero-days including vulnerabilities in OpenBSD (27 years old), FFmpeg (16 years old), and the Linux kernel.
- Project Glasswing launched with 12 named partners: AWS, Apple, Google, Microsoft, NVIDIA, Cisco, CrowdStrike, JPMorgan Chase, Broadcom, Palo Alto Networks, the Linux Foundation, and Anthropic. Another 40+ organizations have access.
- Anthropic committed $100M in usage credits for Glasswing, plus $4M to open-source security ($2.5M to Alpha-Omega/OpenSSF, $1.5M to Apache Software Foundation).
- Fed Chair Jerome Powell and Treasury Secretary Scott Bessent briefed US bank CEOs on Mythos cyber risks the same week.
- Claude Opus 4.7 is the public substitute. Anthropic has conceded it trails Mythos on every major benchmark.
Most weeks in AI, the story is which model shipped. This one is the opposite: a model Anthropic finished, benchmarked, and then decided the public cannot have.
Anthropic announced Claude Mythos Preview. Not long after, it shipped Claude Opus 4.7 as the public flagship. The gap between those two models is the most interesting AI story this year, and almost nobody is writing about it correctly.
What Claude Mythos Actually Is
Mythos is a model Anthropic finished and then pulled from the public launch pipeline. It beats Claude Opus 4.7, GPT-5.4, and Gemini 3.1 Pro on every cybersecurity and advanced coding benchmark the company has published.
The capability claim that matters, in Anthropic's own words on the Project Glasswing page:
Anthropic's stated reason
"AI models have reached a level of coding capability where they can surpass all but the most skilled humans at finding and exploiting software vulnerabilities."
That is not marketing language. Anthropic's system card for Mythos says, bluntly, that releasing it broadly "would give attackers a cornucopia of zero-day exploits for essentially all the software on Earth, including every major operating system and browser."
The specific findings back this up. Mythos has already autonomously discovered thousands of high-severity vulnerabilities across production software, including flaws that had survived a combined 40+ years in OpenBSD, FFmpeg, and the Linux kernel.
Why Anthropic Withheld It
Most coverage is getting this part wrong. The usual framing, "Anthropic built something too dangerous to release," isn't quite right. The sharper version is that Anthropic believes the offensive-defensive balance tips the wrong way if everyone gets Mythos at once.
The thinking runs like this. If a cybersecurity firm with Mythos can find a vulnerability in three hours, so can an attacker with Mythos. The defender then has to patch it before the attacker uses it. Hand both sides the tool at the same time and the attacker wins, because supply chains are slow and patching cycles take weeks.
So Anthropic's bet is to let the defenders run first. Give the Linux Foundation, the Apache Software Foundation, AWS, Apple, Microsoft, Google, and the major security vendors a 90-day head start. Let them find and patch the worst vulnerabilities while the model can't be rented by anyone with a credit card. Then, maybe, expand access.
What Opus 4.7 actually is
The timing of Opus 4.7's release is not a coincidence. Anthropic needed a public flagship that's very good but measurably weaker than Mythos on offensive cyber tasks. Anthropic itself has conceded Opus 4.7 trails Mythos on advanced coding and cybersecurity benchmarks. That gap is the product, not a side effect.
Where Mythos Pulls Ahead
Anthropic has not released a full head-to-head benchmark sheet alongside Mythos. What it has disclosed, combined with UK AISI's independent evaluation, points the same direction: Mythos sits above every shipped model on cyber-offensive tasks.
The clearest public data point comes from AISI. Mythos became the first AI model to clear expert-level capture-the-flag challenges at a 73% pass rate. No prior model had cleared 50% on that tier. It is also the first model to complete AISI's 32-step enterprise-network attack simulation end-to-end, which takes a skilled human red-teamer roughly 20 hours.
Anthropic has separately conceded that Claude Opus 4.7, its public flagship, trails Mythos on advanced coding and cybersecurity benchmarks. The exact spread is not public. What is public is the decision Anthropic made based on it: the partners who need to patch vulnerabilities get Mythos; the rest of us get Opus 4.7.
That gap is also the entire reason Project Glasswing exists.
Inside Project Glasswing
Not sure which AI model to use?
12 models · Personalized picks · 60 seconds
Glasswing is structured as a 90-day defensive-first window. Anthropic published the full partner list and committed to publishing findings within 90 days of launch.
The 12 launch partners:
Launch partners - what each brings
| Partner | Role |
|---|---|
| Anthropic | Hosts Mythos, funds usage credits, reviews findings |
| AWS | Scans its own services; Bedrock integration for enterprise Glasswing members |
| Apple | Scans iOS, macOS, Safari, and first-party services |
| Scans Android, Chrome, Gmail, Google Cloud | |
| Microsoft | Scans Windows, Azure, Office, Edge |
| NVIDIA | Scans driver stack, CUDA, GPU firmware |
| Cisco | Scans networking infrastructure and enterprise routing stack |
| CrowdStrike | Threat intelligence integration and defensive deployment |
| Palo Alto Networks | Firewall and network-security product scans |
| JPMorgan Chase | Financial-services red team, the first major bank in the program |
| Broadcom | VMware, Symantec, and hardware/firmware surfaces |
| Linux Foundation | Open-source kernel, OpenSSF, and Alpha-Omega projects |
Another 40 organizations have access without being named publicly. Anthropic describes them only as entities that "build or maintain critical software infrastructure."
The financial structure matters. Anthropic is paying for the usage credits itself, $100 million committed, rather than charging Glasswing members. That's a notable departure from how every other frontier-AI lab handles high-capability access. On top of that, Anthropic donated $4 million in cash: $2.5M to Alpha-Omega and OpenSSF through the Linux Foundation, and $1.5M to the Apache Software Foundation. Both are organizations that quietly maintain the open-source software stack most of the internet runs on.
The Glasswing bet in one sentence
Anthropic is spending $100M to let 52 organizations find and fix vulnerabilities in the world's critical software before it sells the same capability to anyone who wants it.
What the UK AI Safety Institute Found
The UK AI Safety Institute ran independent evaluations on Mythos Preview and published its findings. Two results matter.
On expert-level capture-the-flag challenges, Mythos cleared 73% of the hardest problems AISI could run against it. No model had cracked 50% on this tier before. Opus 4.6, the prior state of the art, was at a substantially lower pass rate.
On the "Last Ones" (TLO) corporate-network simulation, Mythos did something no model had done before. TLO is a 32-step attack run against a synthetic enterprise network. A skilled human red-teamer takes around 20 hours to complete it. Mythos finished the full chain in 3 out of 10 runs and averaged 22 of 32 steps. Opus 4.6 averaged 16.
AISI's caveat matters here:
AISI's caveat
Test environments "lack active defenders and defensive tooling." AISI "cannot say for sure whether Mythos Preview would be able to attack well-defended systems." The institute's assessment is that Mythos is "capable of autonomously attacking small, weakly defended and vulnerable enterprise systems." Which, to be blunt, covers most small-business networks, most local-government IT, and a disconcerting share of healthcare providers.
Why the Fed and Treasury Are Briefing US Banks
Federal Reserve Chair Jerome Powell and Treasury Secretary Scott Bessent convened the CEOs of major US banks around the time Mythos was announced, according to CNBC and Bloomberg. The topic was Mythos specifically, not AI in general.
The concern isn't that Mythos itself will leak. Anthropic's access controls for Glasswing appear genuinely tight. The concern is about what comes next. If Mythos is the best offensive-cyber AI right now, the second-best is probably months away, not years. Open-weight frontier models from labs outside the Responsible Scaling Policy framework will catch up. US financial regulators are asking whether banks are ready for attacks that run at Mythos-level capability without Anthropic's safety layer.
The answer, judging by the Fed's willingness to hold an emergency briefing, is "not yet." Bloomberg reports that the administration is urging the largest Wall Street banks to voluntarily red-team their own infrastructure against Mythos through Glasswing before any adversary brings an equivalent capability online.
The Precedent This Sets
This is the first time a frontier AI lab has finished a model, benchmarked it at the state of the art, and refused to ship it through normal commercial channels on its own initiative. OpenAI has delayed releases for safety review. Google has gated specific capabilities. No lab had previously said, in public, "we built something, we think it's the best model we've ever trained, and no amount of safety scaffolding is enough to let the public use it."
Two things follow from that.
First, the Responsible Scaling Policy is now load-bearing. Anthropic's RSP framework classifies models into AI Safety Levels (ASL) and was designed for exactly this scenario. Whether Mythos formally crosses an ASL-3 threshold is actually disputed. Some alignment researchers argue it doesn't, and Anthropic withheld it anyway. Either way, the decision sets the precedent that a lab can go beyond its published thresholds if it thinks the situation warrants it. That flexibility cuts both ways.
Second, the public-versus-gated-access debate is no longer theoretical. Anyone who wants frontier cyber-offensive capability now has concrete proof that the capability exists and that Anthropic will not sell it to them. That changes the calculus for every other lab. OpenAI's GPT-5.4-Cyber moved in a similar direction with user-verification gating. Expect more of this, not less.
The open question, and nobody has a good answer to it yet: what happens when the next lab to build a Mythos-level cyber model is not based in a country that runs Responsible Scaling Policies?
FAQ
What's the difference between Claude Mythos and Claude Opus 4.7?
Opus 4.7 is Anthropic's current public flagship. Anyone can use it through the API, Claude.ai, Amazon Bedrock, Google Vertex, or Microsoft Foundry. Mythos is not available publicly. It scores significantly higher on every advanced coding and cybersecurity benchmark. Mythos is available only to roughly 52 vetted organizations through Project Glasswing.
Can my company get Project Glasswing access?
Probably not unless you already build or maintain critical software infrastructure. Anthropic has not published application criteria, but the 40+ additional (unnamed) Glasswing members are reportedly concentrated among major open-source maintainers, critical-infrastructure operators, and financial institutions. Enterprise customers on AWS, Azure, and GCP have access pathways through those cloud partnerships.
Is "Mythos" the final name or a codename?
Anthropic is calling it "Claude Mythos Preview" in official communications, which suggests a successor under the Mythos name may come later. The "Preview" tag is meaningful: it signals an unfinished public strategy rather than a permanent shelf.
Will Claude Mythos ever be released publicly?
Anthropic has not committed to a public release. The Glasswing window is 90 days initially, with a published-findings report due at the end. A broader release depends on what that window surfaces, specifically whether the patching backlog from the zero-days Mythos is finding can be worked through faster than adversaries can build equivalent models.
How does this compare to OpenAI's GPT-5.4-Cyber?
GPT-5.4-Cyber has mandatory user verification for defense and cybersecurity professionals. It is gated but commercially available: customers who pass verification can rent it. Mythos is not rentable at all. Access is by invitation, not transaction. Mythos is also materially more capable on cybersecurity benchmarks.
What should I actually do about this as a non-Glasswing organization?
Three things. First, assume Mythos-level offensive capability will be commercially or illicitly available within 12 months, and price your security posture accordingly. Second, if you run critical software, prioritize reachability reduction and supply-chain hardening now; those are the defenses that work regardless of attacker capability. Third, watch Anthropic's 90-day Glasswing public report. The patching patterns it reveals will telegraph where the remaining soft surfaces are.
Keep Reading
Stay ahead of the AI curve
We test new AI tools every week and share honest results. Join our newsletter.


