How AI-Authored Exploit Code Betrays Itself
The most actionable takeaway for defenders may be the forensic signature itself. GTIG's analysts didn't catch the exploit because of what it did — they caught it because of how it was written. The Python script contained an abundance of educational docstrings, a hallucinated CVSS score for a CVE that didn't yet exist, and a 'structured, textbook Pythonic format highly characteristic of LLMs training data,' including detailed help menus and a clean _C ANSI color class [1]. These are not the habits of a seasoned exploit developer rushing to weaponize a flaw; they are the habits of a model that learned from open-source security tooling and pedagogical examples. The hallucinated CVSS score in particular — a confident, plausible-looking number for a vulnerability the broader world had not yet seen — is the kind of artifact only a generative system produces. For defenders, this opens a new category of detection: 'AI-fingerprint analysis' of suspicious binaries and scripts. For attackers, it implies an emerging operational discipline — stripping pedagogical scaffolding before deployment — that will likely become standard tradecraft within months.



