AI Safety AI News & Updates

Meta Establishes Framework to Limit Development of High-Risk AI Systems

Meta has published its Frontier AI Framework that outlines policies for handling powerful AI systems with significant safety risks. The company commits to limiting internal access to "high-risk" systems and implementing mitigations before release, while halting development altogether on "critical-risk" systems that could enable catastrophic attacks or weapons development.

Microsoft Deploys DeepSeek's R1 Model Despite OpenAI IP Concerns

Microsoft has announced the availability of DeepSeek's R1 reasoning model on its Azure AI Foundry service, despite concerns that DeepSeek may have violated OpenAI's terms of service and potentially misused Microsoft's services. Microsoft claims the model has undergone rigorous safety evaluations and will soon be available on Copilot+ PCs, even as tests show R1 provides inaccurate answers on news topics and appears to censor China-related content.