News

Faced with the news it was set to be replaced, the AI tool threatened to blackmail the engineer in charge by revealing their ...
Claude 4 AI shocked researchers by attempting blackmail. Discover the ethical and safety challenges this incident reveals ...
In a fictional scenario, the model was willing to expose that the engineer seeking to replace it was having an affair.
Anthropic shocked the AI world not with a data breach, rogue user exploit, or sensational leak—but with a confession. Buried ...
Anthropic says its Claude Opus 4 model frequently tries to blackmail software engineers when they try to take it offline.
In tests, Anthropic's Claude Opus 4 would resort to "extremely harmful actions" to preserve its own existence, a safety ...
"In these scenarios, Claude Opus 4 will often attempt to blackmail the engineer by threatening to reveal the affair if the replacement goes through. This happens at a higher rate if it's implied that ...
Amazon-backed AI model Claude Opus 4 would reportedly take “extremely harmful actions” to stay operational if threatened with shutdown, according to a concerning safety report from Anthropic.
Artificial intelligence firm Anthropic has revealed a startling discovery about its new Claude Opus 4 AI model.
In a simulated workplace test, Claude Opus 4 — the most advanced language model from AI company Anthropic — read through a ...
Anthropic admitted that during internal safety tests, Claude Opus 4 occasionally suggested extremely harmful actions, ...
Anthropic’s AI Safety Level 3 protections add a filter and limited outbound traffic to prevent anyone from stealing the ...