News

AI systems like Claude 4 demonstrate significant autonomy, including the ability to identify and report suspicious activities, raising questions about trustworthiness and ethical decision-making.
Investing.com -- Nine days ago, AI lab startup Anthropic released the highly anticipated Opus 4 and Sonnet 4, the next model offerings in the company’s flagship Claude family. A pivotal moment ...
Anthropic’s Claude 4 models, Opus 4 and Sonnet 4, represent significant advancements in AI capabilities, particularly in coding and autonomous task execution. While Opus 4 offers top-tier performance ...
Discover how Anthropic’s Claude 4 AI model is outperforming GPT-4 and Google Gemini with superior coding skills, real-time tool use, long-term memory, and advanced safety protocols. Learn what makes ...
New research shows that as agentic AI becomes more autonomous, it can also become an insider threat, consistently choosing ...
After Claude Opus 4 resorted to blackmail to avoid being shut down, Anthropic tested other models, including GPT 4.1, and ...
New research from Anthropic suggests that most leading AI models exhibit a tendency to blackmail, when it's the last resort in certain tests.
Anthropic’s Claude Opus 4 turned to blackmail 96% of the time, while Google’s Gemini 2.5 Pro had a 95% blackmail rate. OpenAI’s GPT-4.1 blackmailed the executive 80% of the time, and ...
Without proper safeguards, AI could facilitate nuclear and biological threats, among other risks, report commissioned by ...
Major artificial intelligence platforms like ChatGPT, Gemini, Grok, and Claude could be willing to engage in extreme behaviors including blackmail, corporate espionage, and even ...
Claude's distinguishing feature compared to other generative AI models is its focus on "ethical" alignment and safe interactions. On Nov. 11, 2024, Dario Amodei joined Lex Fridman for a 2-and-a ...
Artificial intelligence company Anthropic has released new research claiming that artificial intelligence (AI) models might ...