Back to Blog
March 19, 202610 min read11 views

Claude AI Under Attack: How Chinese Labs Stole Capabilities

claude-aianthropicai-securitydistillationchinese-ai-labsdeepseek

Introduction

In February 2026, Anthropic dropped a bombshell that shook the AI industry: three major Chinese artificial intelligence companies had been running coordinated, industrial-scale campaigns to extract Claude's capabilities and use them to train their own models. The operation involved more than 24,000 fraudulent accounts, over 16 million exchanges with Claude, and sophisticated evasion techniques designed to fly under the radar.

This wasn't a minor terms-of-service violation. It was a systematic effort to reverse-engineer one of the most advanced AI systems in the world. For Claude users, developers building on the API, and anyone following the AI landscape, the implications are significant. This article breaks down exactly what happened, how Anthropic detected it, what distillation actually means in practice, and why this matters for the future of Claude AI and the broader ecosystem.

What Is AI Distillation and Why Does It Matter?

Before diving into the specifics of the attacks, it helps to understand what distillation means in the context of large language models. Distillation is a technique where a smaller or less capable model is trained on the outputs generated by a more powerful AI system. Instead of training from scratch on raw data, the weaker model learns to mimic the reasoning patterns, style, and knowledge embedded in the stronger model's responses.

Think of it like a student copying not just the answers but the entire thought process of a brilliant teacher. Over millions of interactions, the student model begins to replicate capabilities that took the original model enormous resources to develop. The original model's training might have cost hundreds of millions of dollars in compute, data curation, and research. Distillation offers a shortcut that bypasses all of that investment.

This is precisely why distillation attacks are so damaging. They undermine the competitive advantage of frontier AI labs like Anthropic, which invest heavily in safety research, constitutional AI training, and rigorous evaluation. When a competitor can extract those capabilities without bearing any of the cost or responsibility, it distorts the entire market and creates perverse incentives around AI safety.

The Scale of the Operation

The numbers revealed by Anthropic paint a picture of a highly organized, well-funded operation. Three Chinese AI companies were identified as the primary actors: DeepSeek, Moonshot AI, and MiniMax. Together, their campaigns generated over 16 million exchanges with Claude through approximately 24,000 fraudulent accounts.

MiniMax was the most aggressive of the three, driving more than 13 million of those exchanges on its own. To put that in perspective, that volume of interaction represents an enormous amount of structured data about how Claude reasons, responds to edge cases, handles ambiguity, and produces nuanced output across countless domains.

The attacks weren't carried out by individual bad actors poking around with free accounts. These were coordinated campaigns using commercial proxy services to mask the true origin of the requests. In one particularly striking case, a single proxy network managed more than 20,000 fraudulent accounts simultaneously. The operators mixed distillation traffic with unrelated legitimate-looking requests to make the patterns harder to detect, a technique that shows real sophistication in adversarial operations.

All three companies are based in China, where the use of Claude's services is explicitly prohibited due to legal, regulatory, and security concerns. The fact that they circumvented geographic restrictions adds another layer of violation on top of the terms-of-service breaches.

How Anthropic Detected the Attacks

Anthropic has been building detection systems specifically designed to identify distillation attempts, and the discovery of these campaigns demonstrates those systems in action. The company was able to attribute each campaign to a specific AI lab through a combination of technical signals.

IP address correlation played a key role. Even though the attackers used proxy services, patterns in the IP addresses and routing infrastructure provided fingerprints that could be traced back to specific organizations. Request metadata, including the timing, structure, and content patterns of the queries, helped distinguish distillation traffic from normal usage. Infrastructure indicators, such as the specific proxy networks and account creation patterns, provided additional attribution evidence.

What makes this detection noteworthy is that the attackers were actively trying to evade discovery. They mixed their distillation queries with ordinary-looking traffic, rotated through thousands of accounts, and used commercial proxy services rather than obvious VPNs. The fact that Anthropic identified and attributed these campaigns despite those countermeasures suggests that their monitoring capabilities are more advanced than the attackers anticipated.

For Claude API users, this is actually reassuring. It means Anthropic is actively investing in platform integrity and has systems in place to detect abuse patterns that could affect service quality or availability for legitimate users.

The Broader Context: AI Geopolitics and Export Controls

The distillation attacks didn't happen in a vacuum. They came at a time when the United States and China are locked in an increasingly intense competition over AI supremacy, and when debates over AI chip export controls are at the forefront of policy discussions.

Just weeks after Anthropic's disclosure, the company found itself in a separate confrontation with the Pentagon, which designated Anthropic as a supply chain risk to U.S. national security. That designation came after negotiations broke down because Anthropic requested assurances that Claude would not be used for mass surveillance or fully autonomous weapons systems, conditions the Department of Defense was unwilling to accept.

These two stories, the Chinese distillation attacks and the Pentagon dispute, might seem unrelated on the surface, but they illustrate the same fundamental tension. Anthropic is trying to maintain its position as a safety-first AI company while operating in a geopolitical environment where its technology is simultaneously being stolen by foreign competitors and pressured for use by its own government in ways that conflict with its principles.

For the Claude user community, this context matters because it shapes the future trajectory of the platform. Anthropic's willingness to take strong public positions on both fronts, calling out Chinese distillation and pushing back on Pentagon demands, signals a commitment to principled operation that many users value.

What This Means for Claude Users

The immediate impact on everyday Claude users is relatively limited. The distillation attacks didn't compromise user data, didn't degrade Claude's performance, and didn't create any security vulnerability in the consumer or API products. Anthropic detected and shut down the fraudulent accounts, and the company has strengthened its defenses as a result.

However, the longer-term implications are worth considering. First, distillation attacks at this scale could theoretically affect API pricing and availability. When fraudulent accounts consume millions of API calls, that represents real infrastructure cost that ultimately gets distributed across the paying customer base. Anthropic hasn't indicated that pricing was affected, but it's a factor that frontier AI companies will need to account for as these attacks become more common.

Second, the attacks validate the importance of Anthropic's usage policies and geographic restrictions. Some users have criticized the company's decision to block access from certain regions, but the distillation campaigns demonstrate that those restrictions exist for substantive reasons, not arbitrary ones.

Third, for developers building products on the Claude API, the attacks are a reminder of the importance of responsible use and compliance. Anthropic has shown that it has the technical capability to detect and attribute abuse, which means that API users who respect the terms of service have nothing to worry about, while those who don't will eventually be caught.

How Anthropic Is Responding

Beyond detecting and shutting down the specific campaigns, Anthropic has taken several steps to strengthen its defenses against future distillation attempts. The company published a detailed blog post about detecting and preventing distillation attacks, which serves both as a transparency measure and as a deterrent to potential attackers.

Anthropic is also working with industry partners and policymakers to address the systemic issues that enable distillation attacks. This includes collaborating with OpenAI, which has reported similar campaigns from some of the same Chinese AI companies. The fact that multiple frontier AI labs are being targeted suggests this is an industry-wide problem that requires coordinated solutions rather than individual company responses.

On the technical side, Anthropic is investing in more sophisticated detection systems that can identify distillation patterns earlier and with greater accuracy. The company is also exploring watermarking and fingerprinting techniques that could make it easier to prove when a model has been trained on Claude's outputs, even after the distillation process is complete.

The Future of AI Model Protection

The Claude distillation attacks represent a new chapter in AI security, one that goes beyond traditional cybersecurity concerns. Protecting an AI model isn't like protecting a database or a network. The \"product\" being stolen exists in the model's behavior, its reasoning patterns, its ability to handle complex instructions, and its alignment with human values. These properties can be extracted through normal API interactions, which makes them fundamentally harder to protect than static data.

The AI industry is still in the early stages of developing frameworks for model protection. Legal mechanisms like terms of service and copyright law provide some recourse, but enforcement across international borders is challenging. Technical measures like rate limiting, anomaly detection, and output watermarking are helpful but not foolproof. The most effective defense is likely a combination of technical monitoring, legal action, policy advocacy, and industry collaboration.

For Claude users who are also AI developers, this is an emerging area worth watching. As models become more valuable and distillation techniques become more sophisticated, the arms race between model protection and model extraction will intensify. Understanding these dynamics will be increasingly important for anyone building products or services on top of frontier AI systems.

Common Questions About the Distillation Attacks

Many users have raised questions about the attacks since Anthropic's disclosure. One common concern is whether their own Claude conversations could have been affected. The answer is no. The fraudulent accounts operated independently, and Anthropic's detection systems were focused on the attackers' traffic patterns, not on monitoring legitimate user conversations.

Another frequent question is whether the distilled models could become competitors to Claude. In theory, distillation can transfer some capabilities, but it's not a perfect copy. The distilled models would lack the ongoing improvements, safety training, and alignment work that Anthropic continuously applies to Claude. They would also be frozen at the point of extraction, while Claude continues to evolve with each update.

Some users have also asked whether this could lead to stricter API rate limits or additional verification requirements. Anthropic hasn't announced any such changes, and the company appears to be focused on behind-the-scenes detection rather than measures that would burden legitimate users. That said, it's possible that future anti-distillation measures could introduce modest friction for new account creation or high-volume API usage.

Conclusion

The industrial-scale distillation attacks on Claude by Chinese AI labs represent one of the most significant AI security incidents to date. With 24,000 fraudulent accounts, 16 million stolen exchanges, and three major companies implicated, the scope of the operation was remarkable. Anthropic's ability to detect, attribute, and disclose these attacks demonstrates the maturity of its platform security capabilities and its commitment to transparency.

For the Claude community, the key takeaway is that Anthropic takes platform integrity seriously and has the technical sophistication to protect its users and its technology. The broader implications around AI geopolitics, model protection, and industry collaboration will continue to unfold throughout 2026 and beyond.

If you're a power user who wants to stay on top of your Claude usage patterns and ensure you're getting the most out of your subscription, tools like SuperClaude can help you monitor your consumption in real-time and track usage across different models.