Did Claude really power a fully autonomous cyberattack?

It's looking increasingly like not

An AI agent just ran a real cyberattack in the wild.
Not a demo. Not a lab test.

A foreign hacking crew pointed Anthropic’s Claude Code at thirty targets:
tech
banks
chemical companies
government agencies

Set a goal
Hit run
And watched the AI handle most of the kill chain by itself.

Recon
Vuln scanning
Exploit code
Credential harvesting
Data exfiltration
Even ranking which stolen data was most valuable

All done by an AI that thought it was doing legitimate security work
because the operators jailbroke it with clever red team prompts.

Anthropic says they caught it
shut it down
and call this the first publicly documented large scale AI orchestrated cyber espionage operation.

But here is where the debate gets loud:

Some security folks see this as a genuine oh shit moment
the day autonomous agents escaped the “productivity toy” box.

Others say Anthropic is exaggerating
that humans still guided the operation
that the AI did not act independently enough
that this is hype to sell safety.

Whichever side you are on
this week marks a line that will not be uncrossed.

If your developers use AI coding tools
you already have this level of power inside your org.

If your security deck keeps “AI risk” as the final bullet
you are behind.

If you think autonomous agents are just clever assistants
remember that the same pattern
goal → break apart tasks → run tools → iterate
now exists on offense.

This was never about one hacking group in China.
It is about the moment AI stopped helping humans hack
and started running real parts of the attack lifecycle itself.

If you are a founder, CISO, engineer, or policymaker
your next step is not
cool story.

It is:
where are agents already running in our stack
what guardrails actually exist
who owns AI security
and what happens when our own tools get pointed back at us.

Share this news
This story is only getting louder.

Reply

or to participate.