The Model Didn't Get Hacked — The Supply Chain Did
The most important thing to understand about the Mythos incident is that nothing clever happened to the model itself. Anthropic's own investigation statement is unusually precise: 'no evidence that Anthropic's systems are impacted, nor that the reported activity extended beyond the third-party vendor environment.' In other words, the system Anthropic built to protect a tool described as too dangerous to release held. What didn't hold was the perimeter of the companies Anthropic depends on to train and evaluate it.
The attack chain looks roughly like this. Mercor, a $10B AI training-data startup that serves Anthropic, OpenAI, and Meta, was breached by the extortion group Lapsus$, which claimed to have pulled roughly 4TB of data. Inside that dump were technical details about how Anthropic names and hosts unreleased models. A small group operating in a private Discord channel dedicated to tracking unreleased AI models — one member of which is employed by an Anthropic third-party contractor — used what TechCrunch describes as 'an educated guess about the model's online location based on knowledge about the format Anthropic has used for other models' to reach a live Mythos endpoint. That is an access-control and vendor-management story, not an LLM-security story. It is the same failure mode that has quietly produced the last decade of high-profile enterprise breaches, rescripted for the AI supply chain.



