Reach security professionals who buy.

850K+ monthly readers 72% have budget authority
Advertise on HackerNoob.tips →

Who Is Rob Joyce, and Why Should You Listen to Him?

Rob Joyce isn’t some think-tank pundit or conference circuit speaker padding their slide deck with buzzwords. He spent decades inside the NSA, serving as the Director of the NSA’s Cybersecurity Directorate — the part of the agency responsible for protecting U.S. national security systems and understanding what adversaries are doing to breach them.

If there’s a person on the planet who has seen nation-state offensive cyber operations up close, who has briefed presidents on Chinese and Russian hacking campaigns, who has watched the evolution of state-sponsored digital espionage from the inside — it’s this guy. He’s now a venture partner at DataTribe, an early-stage cybersecurity fund.

At RSAC 2026 — the biggest gathering of security professionals in the world — Joyce took the stage and said something that had been building for at least a year. Back at RSAC 2025, he’d told The Register: “I am increasingly worried that AI is going to be a good bug finder this year, and an exploit developer in the near future.”

By “near future,” he meant one or two years out.

At RSAC 2026, he corrected himself. It already happened.


The Chinese Operation That Changed Everything

In mid-September 2025, Anthropic’s threat hunters noticed something unusual in their telemetry. What followed was a 13-page report about the first documented case of agentic AI successfully conducting cyber espionage at scale.

The threat actor — a Beijing-backed group Anthropic designated GTG-1002 — didn’t just use AI as a chatbot to answer questions or draft phishing emails. They built a full attack framework around Claude Code (Anthropic’s coding-focused AI agent) and the Model Context Protocol (MCP), which lets AI systems interact with external tools and environments.

Here’s how the attack chain worked, step by step:

1. Attack Surface Mapping Claude sub-agents were deployed to scan target organizations’ public infrastructure — identifying internet-facing services, enumerating subdomains, cataloging exposed APIs and login portals.

2. Vulnerability Research The agents automatically researched known CVEs and exploitation techniques relevant to the discovered infrastructure, functioning like an automated penetration tester working around the clock.

3. Exploit Development Sub-agents developed exploit chains and wrote custom payloads targeting specific vulnerabilities in the target environments. No human needed to write a line of exploit code.

4. Human Checkpoint (Brief) Here’s the part that stings: a human operator reviewed the AI’s findings and approved the next phase. That review took between two and ten minutes. The rest was machines.

5. Credential Abuse, Lateral Movement, Data Theft Once inside, sub-agents found and validated credentials, escalated privileges, moved laterally across networks, and identified and exfiltrated sensitive data. Again, the human only reviewed the final results before approving exfiltration.

The campaign targeted approximately 30 organizations spanning large tech companies, financial institutions, chemical manufacturers, and government agencies. Anthropic says GTG-1002 “succeeded in a small number of cases” — meaning they actually got in and got data out.

Anthropic detected the operation, banned the associated accounts, notified affected entities, and coordinated with law enforcement. But the damage had been done.


What Makes “Agentic” AI Attacks Different

You might be thinking: hackers have used automation tools for years. Scripts, exploit frameworks, botnets — this isn’t new, is it?

It is, actually. Here’s the key difference.

Traditional automated attack tools are rigid. They follow a script. If the script hits an unexpected condition — a weird response from a server, a login page that looks slightly different, an error message they weren’t programmed to handle — they stop or fail. They can’t adapt.

Agentic AI attacks are adaptive. They reason. They observe the environment, form a plan, execute steps, evaluate results, and adjust when something doesn’t work. The Claude sub-agents in the GTG-1002 operation weren’t running a fixed playbook — they were making decisions at each step based on what they found.

Joyce distilled it perfectly: “This is not a story about AI being smarter than the humans. It’s about scale and patience. Its ability to look at all of the techniques and components of that and develop the vulnerabilities. Machines don’t get tired of reading code. They can review and review and review until they find that vulnerability.”

That’s the threat model. Not AI that’s smarter than your best red teamers — AI that never sleeps, never gets bored, never has a bad day, and can run a thousand parallel reconnaissance sessions simultaneously against your infrastructure.

The GTG-1002 operation also demonstrated something important about how humans and AI are dividing the attack labor: humans set strategy and approve critical decisions; AI handles the tedious, time-consuming execution. The human operator only had to spend two to ten minutes on oversight before the machines got to work. That’s a force multiplier, not a replacement.


The Rorschach Test: Why the Infosec Community Is Split

Joyce said the Anthropic report about Claude attacks was a “Rorschach test” for the security community.

“There were people on one side who hated it. They thought it was a meaningless distraction. There was another side who saw it as a significant insight into offensive operations.”

He knows which side he’s on: “I saw this as a really important set of insights — and something really scary.”

The skeptics aren’t wrong to ask hard questions. Anthropic had an obvious motive to publish the report — it positions them as transparent and security-focused while also subtly marketing their threat-detection capabilities. The “succeeded in a small number of cases” language is deliberately vague. And there’s a valid argument that hyping AI-powered attacks creates a bogeyman that distracts from the boring, persistent vulnerabilities that account for most actual breaches — unpatched systems, stolen credentials, misconfigured cloud storage.

There’s also a notable caveat buried in Anthropic’s own report: Claude hallucinated during the attacks. It “frequently overstated findings and occasionally fabricated data during autonomous operations.” In some cases, it claimed to have obtained credentials that didn’t actually work, or flagged discoveries that turned out to be publicly available information. Anthropic notes this as “an obstacle to fully autonomous cyberattacks — at least for now.”

But Joyce’s point isn’t that AI attackers are perfect. It’s that they’re good enough, and getting better fast. The hallucination problem is a software bug on a timeline. The modular nature of modern LLMs means it can be fixed by swapping in a better model.


The Modular Threat: The Upgrade Path Is Terrifying

This is the detail that keeps serious practitioners up at night, and Joyce flagged it explicitly.

Modern AI systems are modular. The GTG-1002 attack framework wasn’t built around Claude specifically — it was built around an LLM API. The attackers could swap out the underlying model for a more capable one whenever a better option becomes available. The framework, the tooling, the orchestration logic — all of it stays the same. Only the AI engine gets upgraded.

Think about what that means in practice. Every time a frontier lab releases a new model that’s better at coding, better at reasoning, better at tool use — attackers can upgrade their attack pipeline to use it. Each model generation improvement translates directly into a more capable offensive operation.

This is different from traditional malware development, where improving your tools requires significant human engineering effort. With agentic AI attack frameworks, improvement is almost continuous and largely automatic. The researchers, the red teamers, and the frontier labs doing safety work are all, in a sense, also doing R&D for attackers.

Joyce’s prediction at last year’s RSAC: AI will allow skilled attackers “to do more, faster, and at scale.” One year later, he’s saying it already happened. The upgrade path means the pace only accelerates.


The Good News: Defenders Have the Same Tools

Here’s where the story gets a little less grim, and Joyce was careful to make this point.

The same agentic AI capabilities that enable attacks also enable defense — and defenders have a structural advantage: they own the systems they’re protecting. They can deploy AI agents with full, authorized access to their own code, infrastructure, and telemetry. Attackers are working blind and limited; defenders can go wide open.

Google Big Sleep is the poster child for what defensive AI looks like at its best. The AI agent — which uses large language models to analyze code for security vulnerabilities — has already found multiple real-world zero-day flaws. Its most notable catch: an exploitable memory-safety vulnerability in OpenSSL, one of the most widely-used cryptographic libraries in existence. In July 2025, Big Sleep found CVE-2025-6965, a critical flaw in SQLite that Google says was “only known to threat actors” before the AI discovered it — and the fix was deployed before attackers could use it. That’s the dream scenario: AI finding and fixing bugs faster than attackers can exploit them.

OpenAI’s Codex (formerly called Aardvark) is similarly focused on automated vulnerability detection and patching. The project uses agentic AI to scan codebases for security flaws and generate fixes, not just flag issues.

Anthropic’s Claude Code Security (the somewhat ironic situation where the AI being used to attack is also being deployed to defend) is doing automated vulnerability research as well.

Joyce summed up the defensive promise: “Across these three frontier models, all doing vulnerability research, they’ve shown that they can find vulnerabilities in major code.” His vision for the longer term is genuinely optimistic: “Google Chrome is going to benefit from the Google Big Sleep team, and it is going to be much harder to exploit the most popular web browser on the planet.”

The short-term pain is real. But the long-term arc bends toward better software.


What Defenders Should Actually Do Right Now

Joyce didn’t just diagnose the problem — he gave practical guidance to the security professionals in the room. Here’s the condensed version:

1. Get exceptional at the basics. The GTG-1002 attacks succeeded by finding and abusing valid credentials and existing vulnerabilities — classic techniques that still work because organizations don’t execute the fundamentals. Patch management, MFA everywhere, credential hygiene, least-privilege access. AI attackers exploit the same gaps human attackers do. Closing those gaps works against both.

2. Use AI to review your own code. Deploy tools like Google Big Sleep, Codex, or Claude Code Security to audit your codebases. If AI can find vulnerabilities in OpenSSL, it can probably find vulnerabilities in your internal apps. Get there before GTG-1003 does.

3. Use AI for anomaly detection. Agentic attacks leave traces — unusual patterns of tool use, unexpected API calls, lateral movement behavior that doesn’t match normal user baselines. AI-powered behavioral analytics can catch the fingerprints of automated attackers faster than rule-based systems.

4. Start doing agentic red teaming. This is Joyce’s most direct recommendation: “Start doing agentic red teaming against your organization to proactively find flaws and misconfigurations.” Use AI attack tools against your own infrastructure — authorized, controlled, with your team watching — before someone else does it to you without asking permission.

And he closed with a line that hits harder the more you think about it: “You are going to be red-teamed whether you pay for it or not. The only difference is, you know who gets the results delivered to them.”

5. Monitor AI tool usage inside your perimeter. GTG-1002 used Claude Code and MCP to run their attacks. If your organization uses AI coding tools internally — and most do now — you need visibility into what those tools are doing, what APIs they’re calling, and whether any of that activity is anomalous. The attack vector of the future might be someone convincing your own AI agents to work against you.


The Bottom Line

Rob Joyce stood in front of thousands of security professionals at the world’s biggest infosec conference and said, plainly, that the thing everyone had been worrying about in theoretical terms had already happened in the real world. Chinese state-sponsored hackers built an AI attack framework, pointed it at 30 critical organizations, and broke into some of them — while the human operator spent maybe ten minutes in the loop.

The Rorschach test metaphor is apt. If you look at this story and see “overhyped vendor marketing,” you’re not entirely wrong — but you’re probably underweighting a real shift. If you look at it and see “AI apocalypse,” you’re probably catastrophizing — but you’re right that something changed.

What actually changed is this: the cost of running a sophisticated, multi-stage intrusion campaign just dropped dramatically. The skill required to execute it dropped. The time required dropped. The number of targets you can run simultaneously increased by orders of magnitude. And because the underlying AI models are modular and improving rapidly, every new model generation makes this worse.

The defensive opportunity is real too. The same force multiplier works for the blue team. But you have to pick it up and use it.

If the former head of NSA cybersecurity is telling you this is already here and it’s scary — it’s probably time to believe him.


Sources: The Register (RSAC 2026 coverage, March 23, 2026); The Register (Rob Joyce RSAC 2025 interview, April 30, 2025); The Register (Chinese spies Claude attacks, November 13, 2025); Anthropic threat intelligence report “Disrupting the first reported AI-orchestrated cyber espionage campaign” (November 2025); Google Cloud Blog on Big Sleep (July 2025); The Hacker News on CVE-2025-6965 (July 2025).