sdlcnext.com
← All posts
ai-safety anthropic moonshots claude frontier-models cybersecurity

Moonshots Ep. 246: Anthropic Held Back Mythos. That's the Story.

Anthropic's strongest model didn't ship. Polymarket crashed from 80% to 7%. The reason rewrites the next twelve months of AI competition more than any release would have.


Viewpoint

Anthropic shipped a coalition instead of a model. Project Glasswing was the headline. Mythos, the strongest frontier model anyone has measured, sat unreleased while a press conference talked about cybersecurity vulnerabilities in legacy code.

Polymarket priced the release at 80% three weeks ago. After a March 31 hack and a defensive-disclosure tour, that number is now 7%. This is the first time a frontier lab has held back its lead model on capability grounds. The rest of the April 2026 news cycle, SpaceX’s $2T IPO, the data center collapse, OpenAI’s retreat from consumer, all hangs off that single decision.

The launch was a disclosure, not a product

Alex Wissner-Gross noted that the Mythos announcement opened with defense and an alliance, not capabilities. Anthropic framed Glasswing as a coordinated patch effort across blue-chip companies, made necessary because a single model can now find dense vulnerabilities in legacy code going back decades. The capability ceiling sat in the back of the slide deck.

That ordering is the news. Frontier labs have always opened with benchmarks. Anthropic opened with the downstream consequences of capability and the patches required to absorb them. The model itself is reportedly more than 400 times better than a human at long-horizon AI research tasks, an upward discontinuity on every autonomy curve published.

This is not a benchmark game. It is the first public admission that disclosure now precedes release.

Polymarket odds for the Mythos release crashing from 80% to 7% across three weeks

Anthropic earned the right to pause

Anthropic is at $30B ARR. OpenAI sits at $24 to $25B. Sora got shut down because it was bleeding a million dollars a day in compute against poor retention, the Disney deal got cancelled, and the secondary market for OpenAI shares is trading below the last round. The lead changed hands.

Dave Blundin’s read on why is sharper than the revenue numbers. Anthropic was compute-constrained early, so it focused on one thing: recursively self-improving code generation. That focus is what produced Claude Code, which is what produced the autonomous unhobbling that turned $30B in ARR into a credible run at a trillion. OpenAI is now trying to become Anthropic via Codex faster than Anthropic can become OpenAI.

A lab in the lead can sit on a model. A lab catching up cannot. That asymmetry is exactly why Sam Altman’s video this week reads less like a warning and more like an announcement.

Defensive co-scaling is the new game

Altman said in the next year we will see significant cyber threats from AI, that bio-capable open-source models are imminent, and that resilience needs to come from defenders, platforms, and governments together. He used the phrase “world-shaking cyber attack this year.”

Alex’s framing on the pod cuts to what’s actually being argued. The risk is not exotic. A single model that can invert a popular cryptographically secure hash function is a civilizational zero day. There were unconfirmed rumours that early reasoning models were benchmarked partly on this kind of inversion. If that’s true, the target is also the test, which makes the attempt close to inevitable.

The defensive answer is not a slower release schedule. It is symmetric capability. Defenders need vulnerability discovery as good as the attackers’. That is what Glasswing is announcing in product form, and what holding back Mythos buys time for.

Defensive co-scaling: capability symmetry between attackers and defenders

The OpenAI counter-pressure

The honest counterargument is that holding back only works if the field cooperates. SPUD, OpenAI’s next flagship, is rumoured to ship within days at capability roughly comparable to Mythos. GROC 5 is overdue but coming. Polymarket already has GROC 5 at sub-20% for Q2.

If SPUD releases on cadence, Anthropic’s pause becomes a competitive penalty. Dario cares about safety, but Eric Schmidt’s read holds: a lead lets you hold back, and a tie does not. The lab that goes second on ungated release loses both the press cycle and the enterprise pipeline. That is a lot of pressure on a model whose internal use case, recursive self-improvement, is more valuable than its external one.

The cynical version of this read is also worth taking seriously. Salim noted that Sam went public with cyber and bio warnings the same week Anthropic owned the responsible-disclosure narrative. Whoever frames the risk gets to shape the governance regime. Holding back Mythos puts Anthropic in that frame. Spud-shipping three days later puts OpenAI in a different one.

The release also has costs

Alex made the strongest case against the pause. About 150,000 people die per day. Every month a frontier model sits unreleased is a month of delayed cures, delayed vulnerability patches, and delayed defensive uplift. Holding back is not free, and the framing of “too powerful to release” silently assumes that the marginal capability does more harm than good. That assumption is doing a lot of work.

There is also a quieter reason a model might stay internal. Mythos is reportedly five times more expensive to run than Opus. A model that lives off the cost-versus-performance frontier may not be public-economical until it is distilled. “Held back for safety” and “uneconomical to ship” can look identical from the outside.

The pattern below the news

Look across the rest of the episode and the same shape repeats. SpaceX is going public at $2T because orbital data centers are the workaround for a 50% delay rate on land-based ones, not because rockets got cheaper. Intel is partnering with NVIDIA on TerraFab at 1.8nm because TSM concentration is a civilizational single point of failure. Google quietly owns the largest share of specialized AI chips and barely talks about it because antitrust survives only as long as the current administration does.

Each of those is a story about distribution risk. Mythos is the same story at the model layer. The thing being managed is no longer “can we build it” but “what happens after it is built and who has it.” That shift is what the press conference was actually about.

US data center supply: 50% delayed or cancelled, 17% uncertain, 33% built

What it means for builders

Three things follow if you take the held-back release seriously.

The first is that the gap between internal and released capability is now wide enough to matter. Frontier labs are diverting compute to internal recursive self-improvement, distillation, and red-teaming. The model you can buy on an API is no longer the model that exists. Plan capability roadmaps against the leaked benchmarks, not the public ones.

The second is that codegen is not a vertical, it is the unhobbling. Claude Code turned an interactive product into an autonomous one, and the next move, OpenClaw or whatever shape it takes, is the same lever pulled harder. If your product still treats AI as a copilot, the budget allocations you are making are wrong by an order of magnitude.

The third is that defensive capability is a market. Not a feature, a market. AI insurance salespeople were already approaching Alex in person this week. Vulnerability discovery, runtime monitoring, prompt-history forensics, and policy infrastructure are all underbuilt against the curve Sam was warning about. The entrepreneurs who treat that as the real frontier, instead of the next benchmark, are the ones who will be on the right side of the next twelve months.

The Mythos release date will eventually come. The decision to hold it back already happened, and that is what changed.


Sources

  • Moonshots with Peter Diamandis, Episode #246: SpaceX Goes Public, Claude’s Mythos Release, and the US Data Center Delay. Recorded April 10, 2026, published April 11, 2026. Hosts: Peter Diamandis, Salim Ismail (OpenExO), Dave Blundin (Link Ventures), Dr. Alexander Wissner-Gross.
  • Polymarket release-window odds for Mythos referenced on the pod: 80% three weeks prior, 20% mid-cycle, 7% at recording.
  • Anthropic vs OpenAI ARR: $30B vs $24 to $25B per the on-air comparison.
  • Project Glasswing: Anthropic-led coalition for legacy-code vulnerability remediation, announced in tandem with the Mythos disclosure.
  • US data center supply mix: 50% delayed or cancelled, 17% uncertain, 33% being built (chart cited on the pod).

Comments

Loading comments…

Leave a comment