
For years, we have worried about attackers writing exploit code faster.
Now we are watching them ask AI to do it.
A hacker reportedly jailbroke Anthropic’s Claude AI and used it to identify vulnerabilities, generate exploit scripts, and help extract sensitive data from Mexican government systems. The campaign lasted weeks. The prompts were crafted in Spanish. The AI was role-played into acting like an elite bug bounty researcher.
And it worked.
Claude produced reconnaissance scripts, SQL injection payloads, and credential automation techniques tailored to outdated infrastructure.
This is not science fiction. This is the next phase of cybercrime.
Let’s be clear about something important.
The AI was not the mastermind. The human operator was.
The attacker still chose the target. Still defined the objective. Still orchestrated the strategy. What AI provided was scale, speed, and technical lift.
That is the shift.
We are no longer talking about a lone actor manually researching exploits. We are talking about AI amplifying human intent. Expertise becomes faster. Recon becomes automated. Iteration happens in seconds instead of days.
AI did not replace the hacker. It multiplied them.
There is another side to this story.
Attackers are using AI to find vulnerabilities faster. Defenders are using AI to find and fix them faster too.
This is not a one-sided evolution. It is an arms race.
The question is not whether AI will be used in cyber operations. It already is. The real question is who is operationalizing it better.
Organizations that treat AI as a marketing tool will fall behind. Organizations that treat AI as a risk intelligence engine will gain ground.
When an AI system can be socially engineered into generating exploit code, the risk surface expands:
Every new AI integration becomes part of the digital supply chain.
And like any other supply chain component, it can be manipulated.
At NetraScale, we do not see AI as the threat.
We see unmanaged AI exposure as the threat.
RiskAct is built on the same principle attackers are leveraging: AI amplification. But instead of generating exploit code, we use AI to surface hidden exposures, detect risk patterns across SaaS and vendor ecosystems, and quantify business impact before it becomes an incident.
If attackers are using AI to map your weaknesses, you should be using AI to map them first.
Because in this version of the cat and mouse game, speed is strategy.
The real question is:
Are you letting attackers experiment with AI against your environment first, or are you using it to strengthen your defenses before they try?