Anthropic has been privately briefing government officials about Claude Mythos for weeks, telling them the model could enable large-scale cyberattacks. The briefings started before the CMS leak made the model public. That detail matters.
The leaked draft was unusually candid. It admitted that "AI is currently providing more meaningful capability uplift to attackers than to defenders, and that gap is widening." Mythos can chain attack actions autonomously and run multiple hacking campaigns without human oversight. Security analysts at CSO Online noted the model's recursive self-fixing capability, compressing the gap between human and machine software engineering. Between the leaked draft and the external analysis, the picture is of something closer to a weapon than a product.
The question is what you do with that framing. Gizmodo called it directly: the "classic AI company playbook of talking up the dangers of a model to highlight how powerful and capable it is." I think that's right and incomplete at the same time. Anthropic is doing three things simultaneously: fighting the Pentagon over ethical guardrails on military AI use, warning those same officials that its product could facilitate mass cyberattacks, and preparing for what's rumoured to be a $60 billion IPO later this year. Each of those three positions reinforces the other two. The safety brand makes the government warnings credible. The government warnings make the capability story investable. The IPO pressure makes the capability story necessary.
None of which means the warnings are fabricated. Mythos triggered ASL-3 protections under Anthropic's Responsible Scaling Policy, meaning the company's own framework classified it as requiring enhanced security for model weights and deployment restrictions targeting cyber and biological misuse. Whether it approaches ASL-4, the tier defined by models that become "the primary source of national security risk in a major area," hasn't been disclosed. The leaked capabilities suggest the boundary is closer than anyone expected.
I wrote last week about how conveniently the leak landed. The government briefings add another layer. Pre-leak, they look like responsible disclosure. Post-leak, they look like groundwork, the kind of advance positioning that makes an "accidental" revelation feel less accidental. A company that already told the government "this thing is dangerous" has a much easier time controlling the narrative once the public finds out.
CrowdStrike lost roughly $15 billion in market cap on March 27, the day after the leak. Nearly half of cybersecurity professionals now rank agentic AI as their top threat vector. Anthropic gets to sit at the centre of both the problem and the proposed solution, which is a remarkable place to be when you're about to go public.
Sources: