On April 7, 2026, Anthropic published a blog post that most security teams hadn’t fully absorbed by the time it went viral. The headline: an AI model they built — and chose not to release — had independently found thousands of critical vulnerabilities hiding in software that runs the internet, every major operating system, and every major browser. Some of those bugs had been sitting there for decades, surviving millions of automated fuzz tests and years of human review.

The model is called Claude Mythos Preview. The initiative using it is called Project Glasswing. And understanding what both of these mean — not just for Anthropic, but for every organization that depends on software — is quickly becoming a baseline competency for any security leader.

What Anthropic’s Claude Mythos Actually Is

Mythos Preview is Anthropic’s most capable model by a considerable margin — and, crucially, the first frontier AI model any major lab has explicitly withheld from public release because of what it can do. This isn’t a safety decision born of ambiguity. It’s a deliberate choice backed by a stark internal assessment.

According to Anthropic’s own Project Glasswing documentation, Mythos represents a model that is presently far ahead of any other AI in cyber capabilities and presages an era in which AI models can find and exploit vulnerabilities “in ways that far outpace the efforts of defenders.” That language, which appeared in Anthropic’s internal communications before the public announcement, is what triggered stock volatility across major cybersecurity vendors — CrowdStrike, Palo Alto Networks, SentinelOne, and others — when it began circulating in March.

Mythos isn’t a specialized security scanner. It’s a frontier language model whose advanced agentic coding and reasoning capabilities happen to translate with frightening effectiveness into autonomous vulnerability discovery. Give it access to a codebase and a single prompt, and it can identify subtle logic flaws, construct working exploit chains, and document everything — without requiring human steering at each step.

Why this matters beyond cybersecurity: Mythos demonstrates that the gap between “AI that helps you code” and “AI that can systematically break any software it touches” is smaller than the industry assumed. That asymmetry — offense scaling faster than defense — is the core challenge Project Glasswing is trying to answer.

What It Found — And Why That Keeps CISOs Up at Night

The specific vulnerabilities Mythos uncovered aren’t just impressive in aggregate. The type of bugs it found tells you something important about the limits of conventional security tooling.

Consider: a 27-year-old vulnerability in OpenBSD that allowed remote machines to crash. A 16-year-old out-of-bounds write in FFmpeg that automated fuzz testing had touched over 5 million times without flagging. A 17-year-old unauthenticated remote root privilege in FreeBSD (CVE-2026-4747). Multiple Linux kernel vulnerabilities that Mythos chained together to escalate from user-level access to full system control. These weren’t obscure corner cases. They were in widely deployed software that billions of systems depend on.

As Salt Data’s security analysis documented, the FFmpeg finding is particularly instructive. The bug had survived extensive automated testing precisely because discovering it required semantic understanding of intent — what the code was trying to do — not just syntactic pattern matching. Mythos brought that understanding.

“The window between a vulnerability being discovered and being exploited by an adversary has collapsed — what once took months now happens in minutes with AI. Claude Mythos Preview demonstrates what is now possible for defenders at scale, and adversaries will inevitably look to exploit the same capabilities. That is not a reason to slow down; it’s a reason to move together, faster.”

— Elia Zaitsev, CTO, CrowdStrike · Anthropic Project Glasswing blog

The strategic implication Zaitsev is pointing at is the one that should drive your board conversation: the question isn’t whether adversaries will eventually access Mythos-class capabilities. It’s whether your organization will be patched, hardened, and instrumented before they do.

The Benchmarks: Quantifying the Capability Jump

Anthropic published direct benchmark comparisons between Mythos Preview and Claude Opus 4.6 — currently their top publicly available model. The gap is substantial across every relevant dimension.

Benchmark What It Measures Claude Mythos Claude Opus 4.6 Delta
SWE-bench Verified Real-world code bug fixing 93.9% 80.8% +13.1 pts
CyberGym Cybersecurity vuln reproduction 83.1% 66.6% +16.5 pts
Terminal-Bench 2.0 Agentic tool-use in terminal 82.0% 65.4% +16.6 pts
Terminal-Bench 2.1 (extended) Agentic tool-use, longer horizon 92.1%
OSWorld-Verified OS-level interaction tasks 79.6% 72.7% +6.9 pts

Source: Anthropic Project Glasswing announcement, April 2026. All scores represent Mythos Preview at maximum effort with adaptive thinking.

The CyberGym gap (+16.5 points) is the one that matters most for security practitioners. It measures a model’s ability to reproduce known cybersecurity vulnerabilities from documentation — a proxy for how effectively it can understand, replicate, and potentially construct exploit paths. Mythos at 83.1% isn’t just better than Opus 4.6. It’s operating in a different category.

All benchmarks were run internally by Anthropic. No independent replication exists yet, which is a genuine caveat. But the real-world findings — decades-old zero-days in production codebases — function as an external validation that words in a benchmark table can’t fully capture.

Project Glasswing: The Coalition Holding the Keys

Anthropic’s response to having built something it judges too dangerous for public release isn’t to shelve it. It’s to run a structured, gated access program that uses Mythos’ capabilities defensively — finding and patching vulnerabilities in critical infrastructure before adversaries discover them independently.

That program is Project Glasswing. The initial partner coalition includes some of the most significant institutions in global technology and finance:

Amazon Web Services Apple Broadcom Cisco CrowdStrike Google JPMorganChase Linux Foundation Microsoft NVIDIA Palo Alto Networks 40+ Critical Infra Orgs

“We’ve been testing Claude Mythos Preview in our own security operations, applying it to critical codebases, where it’s already helping us strengthen our code. We’re bringing deep security expertise to our partnership with Anthropic and are helping to harden Claude Mythos Preview so even more organizations can advance their most ambitious work with security that sets the standard.”

— Amy Herzog, VP & CISO, Amazon Web Services · Anthropic Glasswing blog

Beyond model access, Anthropic is committing up to $100M in usage credits to Glasswing partners, plus $4M in direct donations — $2.5M to Alpha-Omega and OpenSSF through the Linux Foundation, $1.5M to the Apache Software Foundation — to fund open-source security infrastructure. These donations aren’t symbolic. They fund the maintainer capacity needed to process and patch AI-generated vulnerability reports.

The financial angle matters too. On April 10, 2026 — three days after the Glasswing announcement — CoreWeave and Anthropic announced a multi-year GPU infrastructure agreement to support Claude’s production deployment at scale. CoreWeave reported $5.13B in 2025 revenue with guidance for over $12B in 2026 and a contracted backlog exceeding $66B. Mythos-class workloads don’t run on commodity hardware, and the infrastructure commitments signal that Anthropic is building for sustained operation at frontier scale — not a one-off research demo.

Risk Matrix: What Mythos-Class AI Means for Your Threat Model

If Mythos-class capabilities reach adversaries — whether through model weight leakage, independent development by well-funded state actors, or gradual proliferation as the capability ceiling rises across the industry — the following risks move from theoretical to near-certain. Here’s how to prioritize them.

Critical AI-accelerated zero-day discovery

Models that scan codebases autonomously compress discovery timelines from months to hours. Every major OS and browser is exposed. Patch cycles become the primary survival variable.

Critical Autonomous exploit chaining

Mythos didn’t just find individual bugs — it chained multiple Linux kernel vulnerabilities into a privilege escalation path. AI-driven lateral movement becomes real-time.

High Open-source supply chain exposure

The FFmpeg and OpenBSD findings demonstrate that widely deployed OSS carries latent risk that conventional tooling misses. Every downstream dependency is a potential vector.

High Unmanageable vuln backlogs

If your team can’t patch faster than AI can find and report issues, you’re accumulating disclosed liability. AI discovery without AI-assisted triage creates a new failure mode.

Medium Model access leakage

Gated access programs can leak via insider misuse, prompt extraction, or model weight exfiltration. Current public docs don’t detail Glasswing’s mitigations for this.

Medium Regulatory and compliance friction

Anthropic proactively briefed governments on Mythos’ risks. Central banks and financial regulators are already evaluating systemic cyber-risk implications for large institutions.

The CISO Playbook: 30/90/365-Day Action Framework

You don’t need Mythos access to start hardening for a Mythos-class threat environment. Here’s a sequenced response.

Defensive Framework: AI Zero-Day Era

01

0–30 Days: Assess & Triage

Inventory your critical software, open-source dependencies, and highest-exposure services. Map your current vulnerability discovery pipeline — SAST, DAST, fuzzing, manual review — and identify where semantic understanding gaps exist. These are where Mythos-class systems will find what your tools missed. Check your patch SLA against realistic AI-accelerated exploitation timelines.

02

30–90 Days: Upgrade Detection & Response

Deploy AI-augmented code scanning in your CI/CD pipeline — tools that can reason semantically about code behavior, not just match known patterns. Evaluate whether you qualify for Glasswing-adjacent programs as they expand. Conduct tabletop exercises assuming AI-assisted adversaries. Tune your EDR and XDR stack for novel, AI-generated exploit signatures you’ve never seen in the wild before.

03

90–365 Days: Structural Hardening

Build AI-assisted red team capacity internally or through trusted partners. Implement rigorous SBOM tracking and dependency governance — supply chain exposure was central to Mythos’ most dramatic findings. Establish a board-level reporting cadence for AI cyber risk alongside traditional threat briefings. Push regulators for clarity on AI-generated vuln disclosure obligations before they mandate it.

04

Ongoing: Monitor & Participate

Track Glasswing disclosures and CVE publications linked to AI-discovered vulnerabilities as leading indicators. Participate in ISACs and AI-security working groups. Watch for Anthropic’s planned expansion of Mythos-class access as safeguards mature — the organizations that participated in early access programs historically built the deepest defensive expertise.

“AI capabilities have crossed a threshold that fundamentally changes the urgency required to protect critical infrastructure from cyber threats, and there is no going back. Our foundational work with these models has shown we can identify and fix security vulnerabilities across hardware and software at a pace and scale previously impossible. That is a profound shift, and a clear signal that the old ways of hardening systems are no longer sufficient.”

— Anthony Grieco, SVP & Chief Security & Trust Officer, Cisco · Anthropic Glasswing blog

The Contrarian View: Is This Defense or Theater?

Glasswing’s defenders-first framing has attracted real skepticism, and it deserves engagement rather than dismissal.

The first concern is structural. Concentrating Mythos access in a coalition of Big Tech companies and large financial institutions doesn’t just protect critical infrastructure — it entrenches it. Smaller enterprises, non-US organizations, academic researchers, and open-source communities without Fortune 500 relationships don’t get early access. The Mythos capability gap between Glasswing partners and everyone else may persist for years. Jim Zemlin of the Linux Foundation acknowledged this tension directly, framing Glasswing as a chance to give even resource-constrained maintainers an “AI-powered sidekick.” But early access remains concentrated at the top.

The second concern is practical: disclosure without patch capacity creates liability, not security. Mythos can find vulnerabilities faster than human teams can validate, triage, and fix them. If the AI-generated discovery backlog overwhelms the humans responsible for remediation, the net effect could be a larger disclosed attack surface — not a smaller one.

The third concern cuts to the core of the “defense-first” thesis itself. CrowdStrike’s Zaitsev and Palo Alto’s Lee Klarich both argue that adversaries will develop equivalent capabilities regardless, so the right move is to accelerate defenders. That logic is defensible but not closed. Capable state actors may already have models approaching Mythos-class performance, or they may be years away. The timeline assumption embedded in “move faster together” carries enormous strategic weight — and Anthropic hasn’t published it.

“Perhaps even more important: everyone needs to prepare for AI-assisted attackers.”
— Lee Klarich, Chief Product & Technology Officer, Palo Alto Networks

None of this makes Project Glasswing a bad idea. It makes it an incomplete answer to a problem that will outlast any single initiative. The organizations that treat Glasswing as a complete solution will be wrong. Those who treat it as the opening move in a longer defensive buildout are closer to right.