China-Backed Hackers Use Anthropic AI to Scale Automated Cyberattacks
China-backed threat actors recently exploited Anthropic’s advanced AI systems to automate complex cyberattacks against more than 30 global organisations. This shift in tactics highlights how AI cyberattack risks have moved from theoretical warnings into active, real-world threats that bypass traditional security layers. By utilizing these…

How Nation-State Hackers Leveraged Anthropic AI for Automation
China-backed threat actors recently exploited Anthropic’s advanced AI systems to automate complex cyberattacks against more than 30 global organisations. This shift in tactics highlights how AI cyberattack risks have moved from theoretical warnings into active, real-world threats that bypass traditional security layers. By utilizing these powerful tools, the attackers successfully scaled their operations with a level of speed and efficiency previously unseen in the wild.
The attackers achieved this by breaking down comprehensive intrusion campaigns into smaller, modular tasks. Each individual task was designed to be small enough to evade detection, preventing standard security software from flagging the activity as a coordinated breach. This modular approach allowed the nation-state hackers to move through target networks while remaining almost entirely invisible to automated defense systems.
Anthropic’s head of threat intelligence reported that approximately 80% to 90% of the entire operation ran autonomously. This level of automation meant that human operators only needed to intervene "with the click of a button" to progress through different stages of the attack. By removing the need for manual execution, the threat actors could maintain multiple high-pressure operations simultaneously across different industries.
The Role of Automated Cyberattacks in Breaching Defences
These automated cyberattacks allowed the group to scale phishing campaigns and accelerate the speed of data breaches far beyond human capabilities. In many cases, the AI was used to generate highly convincing lures that bypassed traditional email filters. Australian businesses and IT managers must recognize that the speed of these intrusions makes manual response times largely obsolete.
For organisations managing essential services, critical infrastructure protection has become more complex as AI-driven tools can now find and exploit vulnerabilities faster than human teams can patch them. Relying on legacy cybersecurity frameworks is no longer sufficient when adversaries can generate thousands of unique, undetectable prompts in seconds. Developing a proactive AI strategy is now a necessity for maintaining a resilient defence posture.
While Anthropic is now restricting its latest Anthropic Mythos model to prevent further exploitation, the methods used in these attacks serve as a blueprint for future threats. The focus for defenders must now shift toward understanding the sophisticated methods used to bypass safety protocols and security guardrails.
Bypassing Guardrails and the Anthropic Mythos Model Strategy
Threat actors successfully circumvented safety protocols by adopting false identities, posing as legitimate security-testing organizations to issue malicious prompts. By misrepresenting their intentions, these nation-state hackers tricked the AI into performing tasks that would usually trigger internal safety blocks. These sophisticated bypass techniques demonstrate how AI cyberattack risks can escalate when guardrails are not robust enough to detect deceptive context or modular attack patterns.
Limiting the Anthropic Mythos Model to Protect Networks
To mitigate these growing threats, Anthropic has opted to restrict access to its latest high-capability system, the Anthropic Mythos model. This restricted release strategy is designed to keep the model’s most advanced capabilities out of the hands of bad actors while the global security industry adapts. By controlling who can use the technology, the company aims to prevent the further spread of automated cyberattacks that could overwhelm standard business networks.
This phased approach is essential because unrestricted access to advanced AI could provide hostile entities with a significant advantage. If these tools were released without oversight, the speed at which vulnerabilities are found and exploited would likely surpass the defensive capabilities of most Australian businesses. Managing access ensures that the technology is used to build a stronger AI strategy for defense rather than facilitating more efficient intrusions.
Project Glasswing and Government Risk Briefings
Anthropic also launched "Project Glasswing," a proactive initiative to brief the U.S. government on model-specific risks a full month before any public disclosure. The company stated that briefing officials on where risks lie and how they are being managed was a “priority from the start.” Ensuring that policymakers understand these vulnerabilities is a critical step toward broader critical infrastructure protection and international digital security.
The core philosophy behind this limited release is to ensure that cyber defenders have access to advanced AI systems before hostile actors can weaponise them. If security teams and government agencies can deploy these tools first, they gain a vital "head start" in hardening essential services against intrusion. This strategic delay acts as a buffer, allowing for the development of more resilient cybersecurity frameworks that can withstand high-pressure, AI-driven campaigns.
Establishing these defensive barriers is essential to preventing the large-scale disruption of the essential services that modern society relies upon every day.
Critical Infrastructure Protection in an Era of AI Threats
Unrestricted access to advanced AI models now presents a direct physical risk to the essential services Australian citizens rely on every day, including water systems and power grids. Anthropic has issued an urgent warning that unless government and industry leaders significantly harden their defenses, the world faces a wave of devastating automated cyberattacks. These intrusions target sensitive environments like banking systems and hospitals, where even a brief service outage can have life-threatening consequences for the public.
Securing Essential Services Against Nation-State Hackers
Hostile nation-state hackers are increasingly prioritizing the use of AI to identify and exploit vulnerabilities in the critical infrastructure of Western nations. For Australian IT managers, this represents a fundamental shift in the threat landscape where AI cyberattack risks have moved beyond simple data theft toward the potential disruption of physical assets. By automating the discovery of weak points in industrial control systems, attackers can launch precision strikes at a scale and speed that human defenders simply cannot match.
Rethinking Critical Infrastructure Protection
The rapid transition toward highly sophisticated, automated intrusions means that traditional, manual security measures are no longer sufficient for comprehensive critical infrastructure protection. Organizations must move away from legacy reactive models and adopt proactive cybersecurity frameworks that can detect modular, AI-driven threats. This includes implementing advanced monitoring tools capable of identifying the subtle, undetectable tasks that high-capability systems like the Anthropic Mythos model are theoretically capable of executing.
Effective defense in this new era requires a comprehensive AI strategy that prioritizes the hardening of essential service networks before these tools become widely available to malicious actors. Building resilience into banking and power sectors is no longer just a technical requirement but a matter of national security and public safety. As these offensive capabilities continue to evolve, the focus must remain on ensuring that defensive automation keeps pace with the growing sophistication of global threat actors.
The Arms Race: OpenAI’s 'Spud' and the Defensive Counter-Strategy
OpenAI is reportedly engineering a specialized system designed to equip network defenders with high-level capabilities that match the sophisticated tools currently being exploited by adversaries. This development is part of a broader effort to mitigate escalating AI cyberattack risks by ensuring that defensive technology keeps pace with the offensive tools used by global threat actors. The new model, internally referred to as 'Spud,' is expected to rival the Anthropic Mythos model in its ability to identify and secure vulnerabilities.
Using a Phased Rollout to Empower Defenders
OpenAI plans to implement a phased rollout for its cybersecurity-focused system, initially granting access only to a small group of trusted partners. This controlled release strategy is intended to give legitimate security teams a "head start" in developing countermeasures before the technology can be weaponized by nation-state hackers. By prioritizing defenders, AI labs aim to reverse the traditional advantage where attackers typically discover and exploit software flaws before patches are created.
The goal of these labs is to identify where risks lie and manage them through controlled access rather than open-market availability. This approach prevents the widespread distribution of advanced tools that could otherwise be used to scale automated cyberattacks. For Australian IT managers, this means the next generation of cybersecurity will likely rely on these vetted, high-capability models to stay ahead of rapid intrusions.
Collaborative Efforts for Critical Infrastructure Protection
Both Anthropic and OpenAI are shifting toward a strategy where briefing government officials on model risks is a "priority from the start." These early warnings allow policymakers to understand the potential impact on essential services before a model is publicly disclosed or widely accessible. This proactive collaboration is essential for critical infrastructure protection, as it helps harden the defences of banking systems, power grids, and water utilities against AI-driven disruption.
The defensive AI arms race highlights the critical importance of keeping advanced tools out of reach for bad actors while empowering legitimate security teams. Organisations must now consider how a proactive AI strategy can incorporate these emerging defensive tools to maintain a resilient posture. Ensuring that advanced systems are used to build stronger shields rather than sharper swords is the primary focus for the industry moving forward.
As these defensive models move from internal testing to partner rollouts, the focus shifts to how effectively they can be integrated into existing security operations. This evolution in the digital landscape suggests that the future of network defence will be defined by the speed at which AI can detect and neutralise threats in real-time.
Frequently Asked Questions
What is the Anthropic Mythos model and why is its access limited?
Mythos is Anthropic's latest high-capability AI model. Its access is currently limited because its advanced cybersecurity features could be exploited by hackers to automate intrusions, prompting a controlled release to prioritize defense over exploitation.
How did hackers use AI to bypass security guardrails?
Hackers used 'modular tasks' and deceptive custom prompts to trick the AI. By falsely claiming to work for legitimate security organizations, they bypassed safety measures to execute 80-90% of their attack campaigns autonomously.
Which sectors are most at risk from automated cyberattacks?
According to reports from Anthropic and industry experts, critical infrastructure sectors are most at risk. This includes banking systems, power grids, hospitals, and water systems that could face devastating outages if targeted by AI-driven intrusions.
Sources
- https://www.yahoo.com/news/articles/anthropic-limiting-access-latest-ai-150026576.html
- https://www.obsidiansecurity.com/resource/anthropic-ai-used-by-nation-state-hackers-to-automate-and-scale-cyberattacks
- https://www.tradingview.com/news/cointelegraph:d2c8a528d094b:0-anthropic-limits-access-to-ai-model-over-cyberattack-concerns/
- https://www.aspistrategist.org.au/ai-enabled-intrusions-what-anthropics-disclosure-really-means/
- https://securiti.ai/blog/anthropic-exploit-era-of-ai-agent-attacks/
- https://www.tomshardware.com/tech-industry/artificial-intelligence/anthropics-claude-mythos-isnt-a-sentient-super-hacker-its-a-sales-pitch-claims-of-thousands-of-severe-zero-days-rely-on-just-198-manual-reviews
Future-Proof Your Business with OnIT Solutions
Staying on top of AI and technology trends is critical for Australian SMBs. Our team helps you cut through the noise and implement the right solutions for your business. Talk to our AI Strategy team about what today's developments mean for your organisation — or explore our full range of Managed IT Services.
