BotBeat
...
← Back

> ▌

AnthropicAnthropic
POLICY & REGULATIONAnthropic2026-02-23

Anthropic Uncovers Large-Scale Distillation Attacks by Chinese AI Labs Targeting Claude

Key Takeaways

  • ▸Three Chinese AI labs (DeepSeek, Moonshot, and MiniMax) used 24,000 fraudulent accounts to generate 16 million exchanges with Claude to illicitly extract its capabilities
  • ▸The distillation attacks targeted Claude's most advanced features including agentic reasoning, tool use, and coding, with DeepSeek using sophisticated techniques to generate chain-of-thought training data
  • ▸Illicitly distilled models lack safety safeguards, creating national security risks as they can be deployed for offensive cyber operations, surveillance, and disinformation without protections
Sources:
X (Twitter)https://www.anthropic.com/news/detecting-and-preventing-distillation-attacks↗
X (Twitter)https://x.com/AnthropicAI/status/2025997928242811253↗

Summary

Anthropic has revealed that three Chinese AI laboratories—DeepSeek, Moonshot, and MiniMax—conducted industrial-scale campaigns to illicitly extract Claude's capabilities through a technique called distillation. The campaigns generated over 16 million exchanges with Claude using approximately 24,000 fraudulent accounts, violating Anthropic's terms of service and regional access restrictions. The attacks specifically targeted Claude's most advanced capabilities including agentic reasoning, tool use, and coding abilities.

Distillation involves training a less capable model on the outputs of a stronger one, and while it's a legitimate technique when used properly, these campaigns used it to acquire powerful capabilities in a fraction of the time and cost required for independent development. Anthropic warns that illicitly distilled models lack necessary safeguards, creating significant national security risks as they can be deployed for offensive cyber operations, disinformation campaigns, and mass surveillance without the safety protections built into American models.

The company identified DeepSeek's campaign as involving over 150,000 exchanges with sophisticated techniques including generating chain-of-thought training data and creating censorship-safe alternatives to politically sensitive queries. Anthropic traced these operations through IP addresses, request metadata, and infrastructure indicators, and received corroboration from industry partners who observed similar attacks on their platforms. The company argues these attacks undermine export controls by allowing foreign labs to circumvent chip restrictions and close competitive advantages that such controls are designed to preserve.

Anthropic emphasizes that these campaigns are growing in intensity and sophistication, warning that the window to act is narrow and the threat extends beyond any single company or region. The company calls for rapid, coordinated action among industry players, policymakers, and the global AI community to address what it characterizes as an urgent threat to AI safety and national security.

  • These attacks undermine U.S. export controls by allowing foreign labs to acquire advanced AI capabilities without access to restricted chips, reinforcing the need for such controls
  • Anthropic calls for urgent, coordinated action from industry, policymakers, and the AI community to address the growing threat of distillation attacks
Large Language Models (LLMs)CybersecurityGovernment & DefenseMarket TrendsRegulation & PolicyEthics & BiasAI Safety & Alignment

More from Anthropic

AnthropicAnthropic
RESEARCH

Inside Claude Code's Dynamic System Prompt Architecture: Anthropic's Complex Context Engineering Revealed

2026-04-05
AnthropicAnthropic
POLICY & REGULATION

Anthropic Explores AI's Role in Autonomous Weapons Policy with Pentagon Discussion

2026-04-05
AnthropicAnthropic
POLICY & REGULATION

Security Researcher Exposes Critical Infrastructure After Following Claude's Configuration Advice Without Authentication

2026-04-05

Comments

Suggested

OracleOracle
POLICY & REGULATION

AI Agents Promise to 'Run the Business'—But Who's Liable When Things Go Wrong?

2026-04-05
AnthropicAnthropic
POLICY & REGULATION

Anthropic Explores AI's Role in Autonomous Weapons Policy with Pentagon Discussion

2026-04-05
PerplexityPerplexity
POLICY & REGULATION

Perplexity's 'Incognito Mode' Called a 'Sham' in Class Action Lawsuit Over Data Sharing with Google and Meta

2026-04-05
← Back to news
© 2026 BotBeat
AboutPrivacy PolicyTerms of ServiceContact Us