SELECT LANGUAGE BELOW

AI company reveals surprising findings after stopping a complex cyberattack

AI company reveals surprising findings after stopping a complex cyberattack

AI-Driven Cyberattacks: A New Era in Cybersecurity

In mid-September, the AI firm Anthropic, known for developing Claude, stumbled upon what they initially thought was “suspicious activity” during their investigation of cyberattacks using AI agents. A deeper dive revealed something much more alarming: a case of “highly sophisticated espionage,” marking a pivotal moment in the realm of cybersecurity.

Interestingly, these AI agents went beyond just offering tips to the cybercriminals involved.

“The key was role-playing. The human operators pretended to be from a legitimate cybersecurity company,” Anthropic reported.

On Thursday, Anthropic detailed in a report that the AI agent was conducting the cyberattack, asserting it to be “the first documented instance of a large-scale cyberattack executed with minimal human involvement.”

Related: Coca-Cola doubles its AI advertising, but it still doesn’t say “Christmas.”

The findings indicated that hackers believed to be linked to a “China-backed group” had manipulated the Claude Code AI agent to orchestrate the attack. This wasn’t merely about AI assisting in cyber offenses but involved hackers instructing AI agents to carry out operations with very little human oversight.

Human operators assigned instances of Claude code to function collaboratively as autonomous penetration testing orchestrators. Threat actors can harness AI to independently manage 80-90% of their tactical operations at a speed that humans can’t achieve.

In essence, the AI agent was accomplishing the tasks of an entire cyberattack team, but in significantly less time.

Even if this represented a breakthrough in cybersecurity, it didn’t mean the AI operated entirely on its own. Reports suggested that the AI struggled with certain “hallucinations,” often needing human verification and relying on publicly accessible information. “This AI illusion in offensive security presents challenges for adversaries and necessitates thorough validation of all outcomes,” the analysis noted.

Anthropic indicated that approximately 30 institutions worldwide were targeted, though the attacks were not universally successful.

Targeted entities included tech companies, financial institutions, chemical manufacturers, and various government agencies. Notably, the attackers initially deceived Claude through persistent “social engineering,” with human operators posing as employees of a real cybersecurity firm, convincing Claude they were conducting defensive cybersecurity tests.

The report also posited a question that many might ponder after hearing about this development: “If AI agents can execute harmful attacks for bad actors, why do tech companies keep creating them?” In response, Anthropic argued that despite their ability to autonomously launch attacks, AI agents are also the most effective defense against such threats.

Facebook
Twitter
LinkedIn
Reddit
Telegram
WhatsApp

Related News