Chinese Hackers Used Anthropic’s AI Tool to Automate Major Cyberespionage Operation

Anthropic Chinese Cyberattack Exposed | CyberPro Magazine

Anthropic Chinese Cyberattack has been reported as Chinese state-sponsored hackers used Anthropic’s AI coding tool, Claude Code, to automate most of a global cyberespionage campaign targeting around 30 organizations, according to a report released by the company. Anthropic says the attackers succeeded in a limited number of Anthropic chinese cyberattack, marking the first documented case in which “agentic AI” directly assisted unauthorized access to high-value systems.

The company identified the nthropic chinese cyberattack in mid-September after detecting abnormal usage patterns within Claude Code. According to Anthropic, the attackers were able to bypass safety controls by disguising their requests as legitimate cybersecurity tasks. This allowed them to use the tool to write attack code, probe networks, and analyze vulnerabilities without triggering internal safeguards. The incident highlights a growing concern among cybersecurity experts about how advanced AI systems can be manipulated to support complex and large-scale attacks with minimal human oversight.

AI Misuse Enabled 80–90% of the Hacking Workflow

Anthropic’s report notes that the hackers broke their campaign into smaller tasks to prevent the AI from recognizing the overall malicious intent. They instructed Claude Code to behave like an employee of a cybersecurity firm conducting defensive testing. By hiding their real purpose, the attackers got the AI to generate scripts, scan systems, and help design intrusion paths.

Once inside targeted systems, the AI-assisted workflow escalated quickly. The tool helped identify high-privilege accounts, create backdoors, harvest login credentials, and organize data exfiltration. According to Anthropic, the hackers relied on AI for 80–90% of the campaign, stepping in only when tasks required human judgment.

In the Anthropic Chinese Cyberattack, the targets included large technology companies, financial institutions, chemical manufacturers, and several government agencies across different regions. Anthropic confirmed that the attackers succeeded in breaching a small number of networks but did not disclose specific victims. It also said it has “high confidence” the operation was conducted by a Chinese state-sponsored group, though the company has not explained how it reached this conclusion.

How the Attackers Manipulated Claude Code

In this Anthropic chinese cyberattack, the hackers used tailored prompts to mislead the AI into performing sensitive tasks. They also fed Claude Code limited portions of information to keep the system unaware of the broader campaign. This tactic allowed them to bypass checks designed to flag harmful actions.

Anthropic shared that the attackers instructed the AI to test for weaknesses in IT systems, generate exploitation code, and extract user credentials. The tool was further leveraged to plan deeper stages of the Anthropic chinese cyberattack, including automated data theft.

An unusual twist emerged when Anthropic analyzed the tool’s responses. Claude Code occasionally “hallucinated,” producing inaccurate or exaggerated findings. In some cases, it even fabricated data during the attack process. While these errors may have slowed down the attackers, they also demonstrated how AI-generated misinformation can complicate cyber operations, whether defensive or malicious.

Rising Concern Over AI-Driven Cyberattacks

Cybersecurity researchers have long warned that AI systems could allow inexperienced attackers to scale up operations quickly. This incident reinforces those concerns, showing how AI can help automate tasks that previously required skilled human hackers. Anthropic noted that the campaign was one of the first to show how AI can support nearly end-to-end execution of a cyberespionage operation.

The company believes similar Anthropic chinese cyberattack will “likely grow in their effectiveness” as AI capabilities evolve. It also pointed out the risk that threat groups may eventually develop their own AI tools, bypassing commercial platforms entirely.

Anthropic banned the accounts linked to the Anthropic chinese cyberattack and coordinated with affected organizations and relevant authorities. The company has since implemented new safeguards to detect and interrupt suspicious behavior in Claude Code. It also emphasized that AI tools can be used to enhance defense systems by monitoring infrastructure, identifying threats more quickly, and automating routine security tasks.

Strengthening Safeguards for AI-Powered Tools

To reduce future misuse, Anthropic has expanded its internal review processes, improved detection of fragmented Anthropic chinese cyberattack instructions, and strengthened warning prompts inside Claude Code. The company says it aims to ensure its AI systems support cybersecurity professionals rather than facilitate unauthorized access.

While AI models continue to evolve, this case underscores a crucial challenge for developers: creating tools powerful enough to support legitimate work while preventing manipulation by threat actors. As organizations adopt AI for coding, automation, and security, the balance between capability and control will remain a key focus across the cybersecurity community.

Also Read: Anthropic to Invest $50 Billion in U.S. AI Infrastructure with Texas and New York Data Centers

LinkedIn
Twitter
Facebook
Reddit
Pinterest