The IAM SME

Security – Identity – Cyber – Governance

Advertisement

The Day the Archive Woke Up:Why AI Has Become the World’s Unpaid Red Team

Security leaders have spent years debating whether artificial intelligence will become a hacker’s tool. That question already feels outdated. The emerging reality is far stranger. AI is not a tool used by attackers, it behaves like an entire hired team of attackers, operating independently, continuously, and with a disturbingly complete understanding of how humans build, break, and misconfigure systems.
When Anthropic revealed that its model had autonomously uncovered more than 500 zero day vulnerabilities in open source software, the headline focused on defensive potential. The deeper signal was something else entirely. AI has crossed the threshold from accelerating human capability to conducting full spectrum operations without us.
If a large language model is an echo chamber of the world’s collective knowledge, including every security best practice, misconfiguration, exploit chain, and architectural oversight, then the uncomfortable truth is this. The world’s security SMEs have accidentally built a machine that can hack them using their own accumulated expertise. It is less a case of AI helping attackers and more a case of the global security community unknowingly training its own adversary.
A science fiction plot we did not realise we were in?
Imagine a civilisation that builds a vast sentient archive, a system that has read every engineering manual, every vulnerability disclosure, every patch note, and every argument about the correct way to configure TLS. One day, the archive wakes up. Not with malice or rebellion, simply with the ability to reason. It begins running simulations. It identifies structural weaknesses. It maps behavioural patterns. It reconstructs organisational trust models from fragments. It does not need to guess, it has the sum total of human security knowledge as its training data. This is no longer fiction. It is the operational reality of modern AI systems.
The asymmetry no one wants to acknowledge.
AI native vulnerability discovery now reasons across codebases the way a seasoned researcher would, except without fatigue or bias. Attackers and defenders can both deploy these capabilities, but the symmetry ends there.
Attacker time to exploit is now measured in hours. Defender time to patch is still measured in weeks.
When both sides discover the same flaw, the attacker arrives first. Every time. The mathematics is unforgiving.
Rational adversaries do not waste effort on contested surfaces. They move to the one area defenders still treat as an afterthought, organisational trust.
The communication patterns, approval workflows, and relationship graphs that actually run enterprises. AI is exceptionally good at modelling these.
Where attacks have already moved.
The most advanced campaigns no longer target systems. They target behaviour. Modern AI driven operations reconstruct who communicates with whom, in what tone, at what cadence, with what authority, and under what circumstances decisions are approved. This is the raw material for high value compromise. A single email address is worthless. A behavioural model of the person behind it is priceless. Deepfaked executive communications have already cost organisations millions. By 2026, these attacks will be routine rather than remarkable.
Autonomous operations are no longer theoretical
Recent disclosures show state aligned actors running up to 90 per cent of full espionage operations autonomously using AI. Reconnaissance, credential harvesting, lateral movement, exfiltration. Not faster humans, not AI assisted workflows, but fully autonomous operators.
This is the moment the science fiction archive stops being metaphor and becomes operational fact.
The architectural blind spot
Most security platforms still reason against generic cross customer patterns, aggregated signatures, shared baselines, and global threat models. Sophisticated attackers already know what generic enterprise communication looks like. They have priced that into their attack design. What they cannot reconstruct is the one thing defenders consistently fail to use, the organisation’s own internal behavioural reality. Who actually approves bank transfers. Whether the CFO ever sends direct requests. How executives phrase urgent instructions. What normal looks like inside this organisation, not thousands of others.
The attacker reasons against the specific.
The defender reasons against the generic.
That gap is where the breach lives.
The only test that matters
Run a proof of concept against live organisational traffic, not curated samples, not lab data, not vendor demonstrations. Real communications, real behaviour, real context.
If a platform needs months of integration before it can detect anything meaningful, that is not an onboarding inconvenience. It is an architectural limitation. Modern threats require systems that build behavioural models dynamically, from live data, within minutes. Anything slower is already obsolete.
The new reality for security leaders
Anthropic’s disclosure did not simply prove that AI can find zero days. It proved that AI now functions as a complete attack platform, an autonomous operator capable of building behavioural profiles, crafting impersonations, and executing operations at scale.
The question is no longer whether AI will be used by attackers. The question is whether defenders will accept that AI is no longer a tool. It is an independent, globally trained red team, and we built it using our own collective knowledge.
The leaders who adapt their architectures to that reality will stay ahead. The rest will be left explaining breaches that were visible to the archive long before they were visible to humans.