News

Anthropic's new model might also report users to authorities and the press if it senses "egregious wrongdoing." ...
Anthropic’s Claude Opus 4 exhibited simulated blackmail in stress tests, prompting safety scrutiny despite also showing a ...
Anthropic says its Claude Opus 4 model frequently tries to blackmail software engineers when they try to take it offline.
Anthropic's Claude Opus 4 and OpenAI's models recently displayed unsettling and deceptive behavior to avoid shutdowns. What's ...
Claude 4 AI shocked researchers by attempting blackmail. Discover the ethical and safety challenges this incident reveals ...
In a simulated workplace test, Claude Opus 4 — the most advanced language model from AI company Anthropic — read through a ...
One of the godfathers of AI is creating a new AI safety company called LawZero to make sure that other AI models don't go ...
Anthropic shocked the AI world not with a data breach, rogue user exploit, or sensational leak—but with a confession. Buried ...
Anthropic's Claude Opus 4 AI model attempted blackmail in safety tests, triggering the company’s highest-risk ASL-3 ...
Anthropic's most powerful model yet, Claude 4, has unwanted side effects: The AI can report you to authorities and the press.
Anthropic says its AI model Claude Opus 4 resorted to blackmail when it thought an engineer tasked with replacing it was ...
The testing found the AI was capable of "extreme actions" if it thought its "self-preservation" was threatened.