An AI model just found a security vulnerability that had been hiding for 27 years. One that security experts reviewed for decades, that automated tools tested 5 million times and still missed. And the company that built it decided not to release it.

TL;DR
Claude Mythos built Thousands of zero-days found Too dangerous to release Project Glasswing launched 12 big tech + 40 orgs get access

What Is This?

On April 7, 2026, Anthropic announced Project Glasswing. The story is simple — Anthropic's latest frontier model Claude Mythos Preview has surpassed the best humans at finding and exploiting software vulnerabilities, and they decided not to release it publicly.

Mythos achieved this without any specialized cybersecurity training — purely through improved coding and reasoning abilities. It scored 93.9% on SWE-bench Verified (vs. Opus 4.6's 80.8%) and 77.8% on SWE-bench Pro (vs. 53.4%). As Theo put it — "Mythos is to Opus what Opus is to Sonnet."

The specific results are what's truly alarming. On Firefox exploit generation, where Opus 4.6 managed 2 working exploits out of hundreds of attempts, Mythos hit 181. Here's what it found:

  1. 27-year-old OpenBSD bug
    Found a remote crash vulnerability in one of the world's most security-hardened operating systems. Just connecting to the machine was enough to take it down.
  2. 16-year-old FFmpeg vulnerability
    Caught a flaw in the near-ubiquitous video encoder that automated testing tools had exercised 5 million times without flagging.
  3. Linux kernel privilege escalation chain
    Autonomously found and chained multiple vulnerabilities to escalate from regular user access to complete machine control.
181
Firefox exploits (Opus managed 2)
93.9%
SWE-bench Verified score
$100M
Defensive credits committed

All of this happened fully autonomously, without any human steering. Mythos read the code, found the vulnerabilities, and developed the exploits on its own.

Why Does This Matter?

The key here is an unprecedented decision: an AI company admitting its own model is too dangerous to release. Newton Cheng, Anthropic's Frontier Red Team Cyber Lead, told VentureBeat:

"Given the rate of AI progress, it will not be long before such capabilities proliferate, potentially beyond actors who are committed to deploying them safely. The fallout — for economies, public safety, and national security — could be severe."

— Newton Cheng, Anthropic Frontier Red Team
Traditional AI Security ToolsClaude Mythos
ApproachPattern matching (known vuln DBs)Reasoning-based (understands code context)
AutonomyFollows human-set rulesFully autonomous exploration + exploit dev
Complex attacksDetects individual vulnerabilitiesAuto-chains multiple vulns together
Discovery scopeWithin known patternsIncludes decades-old undiscovered zero-days
AvailabilityAnyone can useRestricted to partners only

Platformer's Casey Newton nailed the tension — this whole initiative "is built on a deeply uncomfortable premise: that the only way to protect us from dangerous AI models is to build them first."

Alex Stamos, CPO at Corridor and former Facebook/Yahoo security chief, gave a blunt warning:

"We only have something like six months before the open-weight models catch up to the foundation models in bug finding. At which point every ransomware actor will be able to find and weaponize bugs without leaving traces for law enforcement to find."

— Alex Stamos, former Facebook security chief

Defenders have months, not years — that's the industry consensus.

Project Glasswing — Arming the Defenders First

Anthropic's response is Project Glasswing. Here's the structure:

  1. Coalition of 12 big tech companies
    AWS, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, Linux Foundation, Microsoft, NVIDIA, and Palo Alto Networks. Competitors Google and Microsoft on the same team tells you how serious this is.
  2. 40+ additional organizations
    Organizations that build or maintain critical software infrastructure get Mythos Preview access to scan and secure both proprietary and open-source systems.
  3. $100M in usage credits + $4M donations
    Anthropic covers the model usage costs for defensive research, plus direct donations to the Linux Foundation and Apache Software Foundation.
  4. Responsible disclosure pipeline
    Vulnerabilities aren't dumped on maintainers. Professional triagers manually validate every report, coordinating at a pace maintainers can handle. 45-day buffer after patches before publishing technical details.

The irony

While Anthropic positions itself as the safe custodian of this unprecedented cyber capability, in March they accidentally leaked 3,000 internal documents through a CMS misconfiguration, and Claude Code's 512,000-line source code was publicly accessible via npm for about 3 hours. Neither was a core system breach, but for a company asking governments to trust it with infrastructure-level security tools, the optics weren't great.

Business Context — Why Now?

On the same day as the Glasswing launch, Anthropic disclosed $30B annualized revenue (3x year-over-year), a 3.5GW compute deal with Google-Broadcom, and the hiring of former Microsoft exec Eric Boyd. Reports suggest an IPO as early as October 2026.

Korean tech outlet THE ELEC analyzed the timing: "beneath the defensive initiative surface lies a complex intent — IPO narrative building, leverage in the Pentagon dispute, and securing partner endorsements."

What's telling is that companies with their own AI security tech — CrowdStrike, Palo Alto Networks — joined the coalition. The interpretation? "They're effectively admitting their own security AI can't defend against Mythos-level attacks."

Key Takeaways: Why This Matters to You

  1. If you build software
    Within 6 months, open-weight models will have similar vulnerability-finding capabilities. Time to overhaul your security pipeline. There's no reason to delay adopting AI security tools.
  2. If you're a business decision-maker
    "Defending AI-created threats with AI" is the new paradigm. Review your security budget and strategy from this perspective.
  3. If you follow the AI industry
    This is the first time an AI company has admitted its own model is too dangerous to release. Not marketing — a new benchmark in AI safety discourse that will directly influence regulation.

The defender's timeline

Anthropic committed to publicly reporting Project Glasswing learnings within 90 days. That's your "defender head start" window. If you're a security practitioner, consider applying to Anthropic's Claude for Open Source program.