AI Ethics AI News & Updates

Pentagon Declares Anthropic National Security Risk Over AI Usage Restrictions

The U.S. Department of Defense has labeled Anthropic an "unacceptable risk to national security" after the AI company imposed restrictions on military use of its technology, specifically refusing uses involving mass surveillance and autonomous lethal targeting. The dispute stems from a $200 million Pentagon contract, with the DOD arguing that Anthropic's self-imposed "red lines" could lead to the company disabling its technology during critical military operations. A court hearing on Anthropic's request for a preliminary injunction against the DOD's designation is scheduled for next week.

Pentagon Develops Independent AI Systems After Anthropic Partnership Collapse

The Pentagon is actively building its own large language models to replace Anthropic's AI following a contract breakdown over military use restrictions. After Anthropic sought contractual clauses prohibiting mass surveillance and autonomous weapons deployment, the Pentagon rejected these terms and instead partnered with OpenAI and xAI. The Department of Defense has designated Anthropic a supply chain risk, effectively barring other defense contractors from working with the company.

AI Industry Rallies Behind Anthropic in Pentagon Supply Chain Risk Designation Dispute

Over 30 employees from OpenAI and Google DeepMind filed an amicus brief supporting Anthropic's lawsuit against the U.S. Department of Defense, which labeled the AI firm a supply chain risk after it refused to allow use of its technology for mass surveillance or autonomous weapons. The Pentagon subsequently signed a deal with OpenAI, prompting industry-wide concern about government overreach and its implications for AI development guardrails. The employees argue that punishing Anthropic for establishing safety boundaries will harm U.S. AI competitiveness and discourage responsible AI development practices.

OpenAI Robotics Lead Resigns Over Pentagon Partnership Citing Governance and Red Line Concerns

Caitlin Kalinowski, OpenAI's robotics lead, resigned in protest of the company's Department of Defense agreement, citing concerns about surveillance of Americans and lethal autonomy without proper guardrails and deliberation. The controversial Pentagon deal, announced after Anthropic's negotiations fell through, has led to a 295% surge in ChatGPT uninstalls and elevated Claude to the top of App Store charts. Kalinowski emphasized her decision was based on governance principles, specifically that the announcement was rushed without adequately defined safeguards.

Anthropic Loses Pentagon Contract Over AI Control Disputes, OpenAI Steps In Despite User Backlash

The Pentagon designated Anthropic as a supply-chain risk after disagreements over military control of AI models for autonomous weapons and mass surveillance use cases. The Department of Defense shifted the $200 million contract to OpenAI, which accepted the terms but experienced a 295% increase in ChatGPT uninstalls afterward. The situation raises questions about appropriate military access to commercial AI systems.

Anthropic's Claude Sees User Surge After Refusing Pentagon Military AI Contract

Anthropic's Claude AI chatbot experienced significant growth in daily active users and app downloads after CEO Dario Amodei refused to allow Pentagon use of Claude for mass surveillance or autonomous weapons, leading to the company being marked as a supply-chain risk. Claude's mobile app downloads now surpass ChatGPT in the U.S., with daily active users reaching 11.3 million on March 2, up 183% from the start of the year. The app reached No. 1 on the U.S. App Store and in 15 other countries, with over 1 million daily sign-ups.

Pentagon Designates Anthropic as Supply Chain Risk Over Refusal to Support Autonomous Weapons and Mass Surveillance

The Department of Defense has officially designated Anthropic as a supply chain risk after CEO Dario Amodei refused to allow military use of its AI systems for mass surveillance of Americans or fully autonomous weapons. This unprecedented designation, typically reserved for foreign adversaries, requires any Pentagon contractor to certify they don't use Anthropic's models, despite Claude currently being deployed in military operations including the Iran campaign. The move has sparked significant criticism from AI industry employees and former government advisors, while OpenAI has signed a deal allowing military use of its systems for "all lawful purposes."

Anthropic's Claude AI Used in US Military Operations Against Iran Despite Corporate Restrictions

Anthropic's Claude AI models are being actively used by the US military for targeting decisions in strikes against Iran, despite President Trump's directive for civilian agencies to discontinue use and plans to wind down DoD operations. Defense contractors like Lockheed Martin are replacing Claude with competitors amid confusion over contradictory government restrictions, while the Pentagon continues using the system with Palantir's Maven for real-time target prioritization. The situation may escalate to a legal battle if the Secretary of Defense officially designates Anthropic as a supply-chain risk.

OpenAI Secures Pentagon AI Contract with Safety Protections Amid Anthropic Standoff

OpenAI has reached an agreement with the Department of Defense to deploy its AI models on classified networks, including technical safeguards against mass domestic surveillance and autonomous weapons. This follows a public conflict between the Pentagon and Anthropic over usage restrictions, which resulted in Trump administration threats to designate Anthropic as a supply-chain risk and ban federal agencies from using its products. OpenAI claims its deal includes protections for the same ethical concerns Anthropic sought, and is asking the government to extend these terms to all AI companies.

Trump Administration Terminates Federal Use of Anthropic AI Following Defense Dispute Over Surveillance and Autonomous Weapons

President Trump ordered all federal agencies to stop using Anthropic products within six months following a dispute with the Department of Defense. The conflict arose when Anthropic refused to allow its AI models to be used for mass domestic surveillance or fully autonomous weapons, positions that Defense Secretary Pete Hegseth deemed too restrictive. Anthropic CEO Dario Amodei maintained the company's stance on these ethical safeguards despite the federal ban.