Latest AI News
OpenAI Finalizes Pentagon Agreement Following Anthropic's Withdrawal
OpenAI announced a deal with the Department of Defense to deploy AI models in classified environments after Anthropic's negotiations with the Pentagon collapsed. The agreement includes stated red lines against mass domestic surveillance, autonomous weapons, and high-stakes automated decisions, though critics question whether the contractual language effectively prevents domestic surveillance. OpenAI defends its multi-layered approach including cloud-only deployment and retained control over safety systems.
Skynet Chance (+0.06%): Deployment of advanced AI models in military classified environments increases potential for dual-use capabilities and loss of civilian oversight, despite stated safeguards. The rushed nature of the deal and ambiguous contractual language around surveillance protections suggest inadequate consideration of alignment and control risks.
Skynet Date (-1 days): Accelerated integration of frontier AI models into military systems shortens the timeline for high-stakes AI deployment with potential control issues. The deal bypasses thorough safety vetting that Anthropic deemed necessary, potentially advancing dangerous applications faster than safety measures can mature.
AGI Progress (+0.01%): The deal primarily concerns deployment contexts rather than capability advances, representing a commercial and regulatory development. While it may provide OpenAI additional resources and data access, it doesn't directly demonstrate progress toward AGI capabilities.
AGI Date (+0 days): Increased Pentagon funding and access to classified use cases could modestly accelerate OpenAI's development resources and real-world testing. However, the primary impact is on deployment rather than fundamental research, yielding minimal timeline acceleration toward AGI.
Trump Administration Blacklists Anthropic Over Refusal to Support Military Surveillance and Autonomous Weapons
The Trump administration has severed ties with Anthropic and invoked national security laws to blacklist the AI company after it refused to allow its technology for mass surveillance of U.S. citizens or autonomous armed drones. MIT physicist Max Tegmark argues that Anthropic and other AI companies have created their own predicament by resisting binding safety regulation while breaking their voluntary safety commitments. The incident highlights the regulatory vacuum in AI development and raises questions about whether other AI companies will stand with Anthropic or compete for the Pentagon contract.
Skynet Chance (+0.04%): The article reveals that major AI companies are abandoning safety commitments and the regulatory vacuum allows development of autonomous weapons systems without safeguards, increasing loss-of-control risks. However, Anthropic's resistance to military applications and the public debate it sparked provide some countervailing pressure against unconstrained AI weaponization.
Skynet Date (-1 days): The competitive pressure created by Anthropic's blacklisting may accelerate other companies' willingness to develop uncontrolled military AI applications, and the abandonment of safety commitments across the industry suggests faster deployment of potentially dangerous systems. The regulatory vacuum means no institutional brakes exist on this acceleration.
AGI Progress (+0.03%): Tegmark's analysis reveals rapid AGI progress, with GPT-4 at 27% and GPT-5 at 57% completion according to rigorous AGI definitions, and AI already achieving gold medal performance at the International Mathematics Olympiad. The article confirms expert predictions from six years ago about human-level language mastery were drastically wrong, indicating faster-than-expected capability growth.
AGI Date (-1 days): The doubling of AGI completion metrics from GPT-4 to GPT-5 in a short timeframe, combined with Tegmark's warning to MIT students that they may not find jobs in four years due to AGI, suggests significant acceleration toward AGI. The competitive dynamics and lack of regulation removing friction from development further accelerate the timeline.