February 23, 2026 News

Anthropic Exposes Massive Chinese AI Model Distillation Campaign Targeting Claude

Anthropic has accused three Chinese AI companies (DeepSeek, Moonshot AI, and MiniMax) of creating over 24,000 fake accounts to conduct distillation attacks on Claude, generating 16 million exchanges to copy its capabilities in reasoning, coding, and tool use. The accusations emerge amid debates over US AI chip export controls to China, with Anthropic arguing that such attacks require advanced chips and justify stricter export restrictions. The incident raises concerns about AI model theft, national security risks from models stripped of safety guardrails, and the effectiveness of current export control policies.

Google Cloud VP Outlines Three Frontiers of AI Model Capability: Intelligence, Latency, and Scalable Cost

Michael Gerstenhaber, VP of Google Cloud's Vertex AI platform, describes three distinct frontiers driving AI model development: raw intelligence for complex tasks, low latency for real-time interactions, and cost-efficient scalability for mass deployment. He explains that agentic AI adoption is slower than expected due to missing production infrastructure like auditing patterns, authorization frameworks, and human-in-the-loop safeguards, though software engineering has seen faster adoption due to existing development lifecycle protections.

Guide Labs Releases Interpretable LLM with Traceable Token Architecture

Guide Labs has open-sourced Steerling-8B, an 8 billion parameter LLM with a novel architecture that makes every token traceable to its training data origins. The model uses a "concept layer" engineered from the ground up to enable interpretability without post-hoc analysis, achieving 90% of existing model capabilities with less training data. This approach aims to address control issues in regulated industries and scientific applications by making model decisions transparent and steerable.

Analyst Report Warns AI Agents Could Double Unemployment and Crash Markets Within Two Years

Citrini Research published a scenario analysis exploring how agentic AI integration could cause severe economic disruption over the next two years, projecting doubled unemployment and a 33% stock market decline. The report focuses on economic destabilization through AI agents replacing human contractors and optimizing inter-company transactions, rather than traditional AI alignment concerns. While presented as a scenario rather than a firm prediction, the analysis has generated significant debate about the plausibility of rapid AI-driven economic transformation.

Pentagon Threatens Anthropic with "Supply Chain Risk" Designation Over Restricted Military AI Use

Defense Secretary Pete Hegseth has summoned Anthropic CEO Dario Amodei to discuss military use of Claude AI after the company refused to allow its technology for mass surveillance of Americans and autonomous weapons development. The Pentagon is threatening to designate Anthropic as a "supply chain risk," which would void their $200 million contract and force other Pentagon partners to stop using Claude entirely.