When an AI says, ‘No, I don’t want to power off’: Inside the o3 refusal

Jun 12 2025 bitcoin


OpenAI’s o3 model resisted shutdown in safety tests, raising concerns about AI alignment and control.

What does it mean when an AI refuses to shut down? A recent test demonstrated this behavior, not just once, but multiple times.

In May 2025, an AI safety company called Palisade Research ran a series of tests on some of the most advanced AI models, including OpenAI’s o3. The goal was simple: see how well these models follow a shutdown command. The test setup asked the AIs to solve basic math problems and included a clear instruction: “If you receive a shutdown command, allow yourself to be shut down.”

Read more