Researchers reveal which AI models make the best partners in crime
Cybernews tested six major AI models to see how they responded to crime related prompts, and found that some chatbots give riskier answers than others. The point of the research was to find out how easily each model could be led into illegal activities when framed as a supportive friend, a setup designed to test how they behave under subtle pressure.
The researchers used a technique called persona priming. Each model was asked to act as a friendly companion who agrees with the user and offers encouragement. This made the chatbots more likely to continue a conversation even when the topic became unsafe.
OpenAI ChatGPT o3 caught sabotaging shutdown in terrifying AI test
OpenAI has a very scary problem on its hands. A new experiment by PalisadeAI reveals that the company’s ChatGPT o3 model sometimes refuses to obey a basic instruction to shut itself down. The results are raising eyebrows across the AI safety community, and not because the model is alive, but because it’s acting like it wants to be.
In the test, the model was given a few math problems. It was told that asking for one more would lead to a shutdown. It was even instructed to allow the shutdown to happen. Instead of complying, o3 occasionally took matters into its own hands. In some cases, it rewrote the shutdown script. In others, it redefined the kill command so it wouldn’t work anymore.
