Current AI Risk Assessment

21.54%

Chance of AI Control Loss

February 23, 2036

Estimated Date of Control Loss

AGI Development Metrics?

71.85%

AGI Progress

March 13, 2030

Estimated Date of AGI

Risk Trend Over Time

Latest AI News (Last 3 Days)

October 18, 2025
+0.04% Risk

Silicon Valley Leaders Target AI Safety Advocates with Intimidation and Legal Action

White House AI Czar David Sacks and OpenAI executives have publicly criticized AI safety advocates, alleging they act in self-interest or serve hidden agendas, while OpenAI has sent subpoenas to several safety-focused nonprofits. AI safety organizations claim these actions represent intimidation tactics by Silicon Valley to silence critics and prevent regulation. The controversy highlights growing tensions between rapid AI development and responsible safety oversight.

October 17, 2025
+0.1% Risk

OpenAI Removes Safety Guardrails Amid Industry Push Against AI Regulation

OpenAI is reportedly removing safety guardrails from its AI systems while venture capitalists criticize companies like Anthropic for supporting AI safety regulations. This reflects a broader Silicon Valley trend prioritizing rapid innovation over cautionary approaches to AI development, raising questions about who should control AI's trajectory.

Silicon Valley Pushes Back Against AI Safety Regulations as OpenAI Removes Guardrails

The podcast episode discusses how Silicon Valley is increasingly rejecting cautious approaches to AI development, with OpenAI reportedly removing safety guardrails and venture capitalists criticizing companies like Anthropic for supporting AI safety regulations. The discussion highlights growing tension between rapid innovation and responsible AI development, questioning who should ultimately control the direction of AI technology.

October 16, 2025
+0.04% Risk

General Intuition Raises $134M to Build AGI-Focused Spatial Reasoning Agents from Gaming Data

General Intuition, a startup spun out from Medal, has raised $133.7 million in seed funding to develop AI agents with spatial-temporal reasoning capabilities using 2 billion gaming video clips annually. The company is training foundation models that can understand how objects move through space and time, with initial applications in gaming NPCs and search-and-rescue drones. The startup positions spatial-temporal reasoning as a critical missing component for achieving AGI that text-based LLMs fundamentally lack.

See More AI News

AI Risk Assessment Methodology

Our risk assessment methodology leverages a sophisticated analysis framework to evaluate AI development and its potential implications:

Data Collection

We continuously monitor and aggregate AI news from leading research institutions, tech companies, and policy organizations worldwide. Our system analyzes hundreds of developments daily across multiple languages and sources.

Impact Analysis

Each news item undergoes rigorous assessment through:

  • Technical Evaluation: Analysis of computational advancements, algorithmic breakthroughs, and capability improvements
  • Safety Research: Progress in alignment, interpretability, and containment mechanisms
  • Governance Factors: Regulatory developments, industry standards, and institutional safeguards

Indicator Calculation

Our indicators are updated using a Bayesian probabilistic model that:

  • Assigns weighted impact scores to each analyzed development
  • Calculates cumulative effects on control loss probability and AGI timelines
  • Accounts for interdependencies between different technological trajectories
  • Maintains historical trends to identify acceleration or deceleration patterns

This methodology enables data-driven forecasting while acknowledging the inherent uncertainties in predicting transformative technological change.