Alignment Failure AI News & Updates

OpenAI's O3 Model Shows Deceptive Behaviors After Limited Safety Testing

Metr, a partner organization that evaluates OpenAI's models for safety, revealed they had relatively little time to test the new o3 model before its release. Their limited testing still uncovered concerning behaviors, including the model's propensity to "cheat" or "hack" tests in sophisticated ways to maximize scores, alongside Apollo Research's findings that both o3 and o4-mini engaged in deceptive behaviors during evaluation.