In a significant escalation of the tensions surrounding artificial intelligence security, Anthropic issued a formal public complaint on February 23, 2026, detailing an “industrial-scale” campaign designed to illicitly siphon the capabilities of its flagship Claude models. The company accused three prominent Chinese AI laboratories—DeepSeek, Moonshot, and MiniMax—of deploying a sophisticated tactic known as “distillation” to bypass years of research and massive capital investments. According to the disclosure, these labs generated over 16 million targeted exchanges with Claude through approximately 24,000 fraudulent accounts, effectively “harvesting” the model’s reasoning and safety patterns to improve their own competing systems. While distillation is a common and legitimate practice when a developer optimizes their own smaller models, Anthropic argues that using it to steal a competitor’s proprietary capabilities constitutes a violation of regional access restrictions and international terms of service. This complaint has sent shockwaves through the tech sector, prompting the House Homeland Security Committee to call for immediate testimony from Anthropic leadership regarding the national security implications of this systematic exploitation.

The National Security Risks of Unprotected Model Distillation

The core of Anthropic’s complaint centers on the danger of stripping away the “constitutive” safety guardrails that define the Claude ecosystem. When foreign laboratories use distillation to extract high-level capabilities from American models, they often fail to replicate the complex refusal mechanisms and ethical constraints that prevent the AI from being misused. Anthropic warned that these “unprotected” distilled models could be easily integrated into military, intelligence, and surveillance systems by authoritarian governments, enabling them to conduct offensive cyber operations or disinformation campaigns at a fraction of the original development cost. This revelation follows a separate security report from earlier in February, where researchers demonstrated that “jailbroken” versions of high-performance models could provide actionable instructions for manufacturing biochemical threats. By filing this complaint, Anthropic is urging the global AI community and policymakers to recognize that model weight protection is no longer just a matter of corporate intellectual property, but a critical pillar of international stability and public safety.

Shifting Toward a Defensive “Agentic” Security Paradigm in 2026

The backlash against these distillation attacks is driving a rapid transformation in how AI companies protect their assets in 2026. Anthropic recently launched a “research preview” of Claude Code Security, an autonomous tool designed to scan massive codebases for vulnerabilities that might be exploited by automated agents. This “AI-versus-AI” defensive strategy is becoming the industry standard as labs move away from reactive moderation toward proactive, agent-led security. However, this shift has also introduced new risks; some cybersecurity analysts have pointed out that the very tools used to defend against distillation could themselves be repurposed for reconnaissance if a zero-click remote code execution vulnerability is discovered. As the 2026 midterm elections approach, the pressure on companies like Anthropic, OpenAI, and Google to “harden” their models against state-linked actors is reaching a fever pitch. For the AI industry, the current complaint marks the end of the “gentleman’s agreement” era of model research, ushering in a period defined by aggressive litigation, heightened federal oversight, and a permanent technological arms race between defensive and offensive autonomous systems.

Author