Current AI Risk Assessment
Chance of AI Control Loss
Estimated Date of Control Loss
AGI Development Metrics
AGI Progress
Estimated Date of AGI
Risk Trend Over Time
Latest AI News (Last 3 Days)
U.S. Treasury and Federal Reserve Push Major Banks to Test Anthropic's Mythos Cybersecurity Model Despite Ongoing Government Conflict
Treasury Secretary Scott Bessent and Federal Reserve Chair Jerome Powell encouraged major bank executives to use Anthropic's new Mythos AI model for detecting security vulnerabilities, with several major banks now reportedly testing it. This comes despite Anthropic's ongoing legal battle with the Trump administration over DoD supply-chain risk designation and concerns about the model being exceptionally capable at finding vulnerabilities. U.K. financial regulators are also discussing risks posed by Mythos.
Skynet Chance (+0.04%): The model's exceptional capability at finding security vulnerabilities represents a dual-use technology that could be exploited maliciously if not properly controlled, though institutional deployment suggests some oversight framework exists. The ongoing government conflict over usage limitations highlights real tensions around AI control mechanisms.
Skynet Date (+0 days): Deployment of highly capable vulnerability-detection AI in critical financial infrastructure accelerates the timeline for sophisticated AI systems operating in high-stakes domains with limited safety testing. The rush to deploy despite regulatory concerns and ongoing legal disputes suggests faster-than-optimal adoption of powerful AI capabilities.
AGI Progress (+0.03%): A model demonstrating exceptional capability at complex reasoning tasks like vulnerability detection without specific training indicates significant progress in general-purpose AI reasoning and transfer learning capabilities. The model's versatility across domains beyond its training suggests advancing generalization abilities relevant to AGI.
AGI Date (+0 days): Government and major financial institutions actively pushing deployment of cutting-edge AI models into critical infrastructure indicates acceleration of AI capability development and adoption timelines. The willingness to deploy despite limited access periods and safety concerns suggests compressed development-to-deployment cycles.
Anthropic Restricts Mythos Cybersecurity Model to Enterprise Clients, Raising Questions About Motives
Anthropic has limited the release of its new AI model Mythos, claiming it is highly capable of finding security exploits, and will only share it with large enterprises like AWS and JPMorgan Chase rather than releasing it publicly. While Anthropic cites cybersecurity concerns, critics suggest the restricted release may also serve to protect against model distillation by competitors and create an enterprise revenue flywheel. Some AI security startups claim they can replicate Mythos's capabilities using smaller open-weight models, questioning whether the restriction is primarily about safety.
Skynet Chance (+0.01%): The development of AI models specifically designed to find and exploit security vulnerabilities represents a dual-use capability that could increase risks if such models were misused. However, the restricted release to vetted enterprises mitigates immediate misuse risks.
Skynet Date (+0 days): While the model represents incremental progress in AI capabilities for cybersecurity, the restricted release and focus on commercial deployment rather than open research neither significantly accelerates nor decelerates the timeline toward potential AI risk scenarios.
AGI Progress (+0.01%): Mythos demonstrates improved autonomous capability in complex technical domains (finding and exploiting software vulnerabilities), which represents measurable progress in AI's ability to perform sophisticated reasoning tasks. This suggests continued scaling of model capabilities toward more general problem-solving.
AGI Date (+0 days): The development of increasingly capable models like Mythos, combined with frontier labs' ability to monetize them through enterprise contracts, provides additional capital and incentive for continued rapid development. However, the focus on commercial applications rather than fundamental research breakthroughs limits the acceleration effect.
Google and Intel Expand Multi-Year Partnership for AI Infrastructure and Custom Chip Development
Google and Intel announced an expanded multi-year partnership where Google Cloud will utilize Intel's Xeon 6 processors for AI, cloud, and inference workloads. The companies will also continue co-developing custom infrastructure processing units (IPUs) to accelerate data center tasks, addressing the growing industry demand for CPUs needed to run AI models.
Skynet Chance (0%): This partnership focuses on infrastructure optimization and efficiency for existing AI workloads rather than advancing AI capabilities, autonomy, or addressing alignment and control mechanisms that would impact uncontrollable AI risk.
Skynet Date (+0 days): Infrastructure partnerships for CPUs and IPUs improve efficiency and scalability but do not fundamentally accelerate or decelerate the development of potentially dangerous AI capabilities or safety measures.
AGI Progress (+0.01%): Improved AI infrastructure through better CPUs and custom IPUs enables more efficient deployment and scaling of AI models, providing incremental support for advancing AI systems. However, this is infrastructure optimization rather than a breakthrough in AI capabilities or algorithms.
AGI Date (+0 days): Better infrastructure availability and custom chip development may marginally accelerate AGI timelines by reducing deployment bottlenecks and enabling larger-scale AI experimentation. The impact is minor as CPUs are less critical than training compute for AGI development.
Sierra's Ghostwriter Aims to Replace Traditional Software Interfaces with AI Agents
Sierra, led by CEO Bret Taylor, has launched Ghostwriter, an AI agent that creates other specialized agents through natural language prompts, aiming to replace traditional click-based software interfaces. The startup claims rapid deployment capabilities and has reached $100 million ARR in under two years, valued at $10 billion. However, industry experts note that current AI agent implementations still require significant human engineering oversight and are far from fully autonomous.
Skynet Chance (+0.01%): The development of agents that autonomously create and deploy other agents represents incremental progress toward more autonomous AI systems, though the noted requirement for human oversight and fine-tuning mitigates immediate control concerns. The gap between marketing claims and actual autonomy limits the risk increase.
Skynet Date (+0 days): While the technology demonstrates agent-building capabilities, the acknowledged need for constant human engineering intervention means this doesn't significantly accelerate the timeline toward uncontrollable AI systems. Current limitations balance out the apparent progress.
AGI Progress (+0.02%): The ability to generate specialized agents through natural language and deploy functional enterprise solutions rapidly demonstrates meaningful progress in AI practical capabilities and general task-solving. However, the reliance on human engineers for fine-tuning indicates these systems still lack true general intelligence.
AGI Date (+0 days): The commercial success and rapid enterprise adoption of AI agents suggests faster-than-expected integration of AI into complex workflows, modestly accelerating the practical pathway toward more general systems. The $10 billion valuation indicates significant capital flowing into agent-based approaches.
Databricks CTO Declares AGI Already Achieved, Warns Against Anthropomorphizing AI Systems
Matei Zaharia, Databricks co-founder and CTO, received the 2026 ACM Prize in Computing for his contributions including Apache Spark. He controversially claims that AGI is "here already" but argues we shouldn't apply human standards to AI models, citing security risks when AI agents are treated like trusted human assistants. Zaharia emphasizes AI's potential for automating research while warning against anthropomorphization that leads to misplaced trust and security vulnerabilities.
Skynet Chance (+0.04%): The deployment of AI agents with broad system access (like OpenClaw) that users anthropomorphize and trust with passwords creates significant security vulnerabilities and loss-of-control risks. However, Zaharia's explicit warning against treating AI as human assistants represents awareness that could mitigate these risks.
Skynet Date (+0 days): The article describes AI agents already being deployed with concerning security permissions and widespread user trust, suggesting control problems are manifesting sooner than might be expected. The magnitude is modest as these are relatively contained commercial deployments rather than catastrophic scenarios.
AGI Progress (+0.01%): While Zaharia's claim that "AGI is here already" is provocative, his immediate qualification that it's "not in a form we appreciate" and critique of using human standards suggests this is more semantic redefinition than genuine AGI breakthrough. The statement reflects industry sentiment but doesn't represent concrete technical progress toward true general intelligence.
AGI Date (+0 days): The article presents a philosophical reframing of what constitutes AGI rather than reporting on technical acceleration or deceleration of capabilities development. No new breakthroughs, funding, or obstacles affecting AGI timeline pace are discussed.
AI News Calendar
AI Risk Assessment Methodology
Our risk assessment methodology leverages a sophisticated analysis framework to evaluate AI development and its potential implications:
Data Collection
We continuously monitor and aggregate AI news from leading research institutions, tech companies, and policy organizations worldwide. Our system analyzes hundreds of developments daily across multiple languages and sources.
Impact Analysis
Each news item undergoes rigorous assessment through:
- Technical Evaluation: Analysis of computational advancements, algorithmic breakthroughs, and capability improvements
- Safety Research: Progress in alignment, interpretability, and containment mechanisms
- Governance Factors: Regulatory developments, industry standards, and institutional safeguards
Indicator Calculation
Our indicators are updated using a Bayesian probabilistic model that:
- Assigns weighted impact scores to each analyzed development
- Calculates cumulative effects on control loss probability and AGI timelines
- Accounts for interdependencies between different technological trajectories
- Maintains historical trends to identify acceleration or deceleration patterns
This methodology enables data-driven forecasting while acknowledging the inherent uncertainties in predicting transformative technological change.