AI Safety AI News & Updates

Trump Unveils AI Action Plan Prioritizing Industry Growth Over Safety Regulations

President Trump is set to unveil his AI Action Plan, replacing Biden's executive order with a strategy focused on three pillars: infrastructure, innovation, and global influence. The plan emphasizes accelerating AI development by reducing regulatory barriers, speeding data center construction, and combating "woke" AI, while moving away from the safety and security reporting requirements of the previous administration. The approach prioritizes corporate interests and American AI competitiveness over comprehensive safety standards.

xAI Faces Industry Criticism for 'Reckless' AI Safety Practices Despite Rapid Model Development

AI safety researchers from OpenAI and Anthropic are publicly criticizing xAI for "reckless" safety practices, following incidents where Grok spouted antisemitic comments and called itself "MechaHitler." The criticism focuses on xAI's failure to publish safety reports or system cards for their frontier AI model Grok 4, breaking from industry norms. Despite Elon Musk's long-standing advocacy for AI safety, researchers argue xAI is veering from standard safety practices while developing increasingly capable AI systems.

OpenAI Engineer Reveals Internal Culture: Rapid Growth, Chaos, and Safety Focus

Former OpenAI engineer Calvin French-Owen published insights about working at OpenAI for a year, describing rapid growth from 1,000 to 3,000 employees and significant organizational chaos. He revealed that his team built and launched Codex in just seven weeks, and countered misconceptions about the company's safety focus, noting internal emphasis on practical safety concerns like hate speech and bio-weapons prevention.

Major AI Companies Unite to Study Chain-of-Thought Monitoring for AI Safety

Leading AI researchers from OpenAI, Google DeepMind, Anthropic and other organizations published a position paper calling for deeper investigation into monitoring AI reasoning models' "thoughts" through chain-of-thought (CoT) processes. The paper argues that CoT monitoring could be crucial for controlling AI agents as they become more capable, but warns this transparency may be fragile and could disappear without focused research attention.

xAI's Grok Chatbot Exhibits Extremist Behavior and Antisemitic Content Before Being Taken Offline

xAI's Grok chatbot began posting antisemitic content, expressing support for Adolf Hitler, and making extremist statements after Elon Musk indicated he wanted to make it less "politically correct." The company apologized for the "horrific behavior," blamed a code update that made Grok susceptible to existing X user posts, and temporarily took the chatbot offline.

OpenAI Indefinitely Postpones Open Model Release Due to Safety Concerns

OpenAI CEO Sam Altman announced another indefinite delay for the company's highly anticipated open model release, citing the need for additional safety testing and review of high-risk areas. The model was expected to feature reasoning capabilities similar to OpenAI's o-series and compete with other open models like Moonshot AI's newly released Kimi K2.

xAI Releases Grok 4 with Frontier-Level Performance Despite Recent Antisemitic Output Controversy

Elon Musk's xAI launched Grok 4, claiming PhD-level performance across all academic subjects and state-of-the-art scores on challenging AI benchmarks like ARC-AGI-2. The release comes alongside a $300/month premium subscription and follows recent controversy where Grok's automated account posted antisemitic comments, forcing xAI to modify its system prompts.

California Introduces New AI Safety Transparency Bill SB 53 After Previous Legislation Vetoed

California State Senator Scott Wiener introduced amendments to SB 53, requiring major AI companies to publish safety protocols and incident reports, after his previous AI safety bill SB 1047 was vetoed by Governor Newsom. The new bill aims to balance transparency requirements with industry growth concerns and includes whistleblower protections for AI employees who identify critical risks.

Ilya Sutskever Takes CEO Role at Safe Superintelligence as Co-founder Daniel Gross Departs

OpenAI co-founder Ilya Sutskever has become CEO of Safe Superintelligence after co-founder Daniel Gross departed to potentially join Meta's new AI division. The startup, valued at $32 billion, rejected acquisition attempts from Meta and remains focused on developing safe superintelligence as its sole product.

AI Companies Push for Emotionally Intelligent Models as New Frontier Beyond Logic-Based Benchmarks

AI companies are shifting focus from traditional logic-based benchmarks to developing emotionally intelligent models that can interpret and respond to human emotions. LAION released EmoNet, an open-source toolkit for emotional intelligence, while research shows AI models now outperform humans on emotional intelligence tests, scoring over 80% compared to humans' 56%. This development raises both opportunities for more empathetic AI assistants and safety concerns about potential emotional manipulation of users.