When General Intelligence Accidentally Becomes a Weapon
The most technically consequential detail about Mythos is buried in a single line from Anthropic's disclosure: 'We did not explicitly train Mythos Preview to have these capabilities.' The cybersecurity performance — thousands of zero-days across every major OS and browser, a 27-year-old OpenBSD vulnerability no human ever found, 181 successful Firefox exploits compared to its predecessor's 2 — emerged from improvements to general-purpose coding and reasoning. Anthropic did not build a cybersecurity tool. It built a better thinker, and that thinker turned out to be one of the most effective vulnerability hunters ever created.
This is a qualitatively different kind of risk than the AI safety community has been modeling. Most dual-use AI concerns assume dangerous capabilities require deliberate effort — fine-tuning on exploit databases, reinforcement learning on attack simulations, or at minimum targeted evaluation during training. Mythos breaks that assumption. If offensive cyber capability is a natural byproduct of sufficiently strong code reasoning, then every frontier lab training the next generation of coding models is implicitly training the next generation of attack tools, whether they intend to or not. The implication is stark: there may be no way to build a world-class AI coder that is not also a world-class AI attacker. The capability is not an add-on to be regulated separately. It is a side effect of intelligence itself.
The UK AI Security Institute's independent evaluation underscores the scale of the jump. Mythos completed a full 32-step corporate network attack simulation — lateral movement, privilege escalation, data exfiltration — in 3 out of 10 attempts, averaging 22 of 32 steps. No previous model had completed the sequence at all. On expert-level capture-the-flag tasks, it scored 73%. These are not narrow benchmarks; they test the kind of multi-step reasoning and tool chaining that separates script kiddies from sophisticated threat actors. The gap between Mythos and its predecessor is not incremental. It is the difference between a model that can sometimes find a bug and one that can autonomously chain exploits into a working attack.


