Category
Anthropic
Claude models, Constitutional AI, and safety research.
7 articles

Anthropic Discovers Functional Emotion Representations Inside Claude, Publishes Landmark Research
New research from Anthropic reveals that Claude develops internal representations resembling emotions — not as experience, but as functional states that influence model behavior in measurable ways.

Federal Judge Blocks Trump Administration's Ban on Anthropic, Calls It 'First Amendment Retaliation'
A federal judge issues a preliminary injunction blocking the Trump administration from banning Anthropic's Claude AI from government use, ruling the move was illegal retaliation for the company's public stance on autonomous weapons.

Anthropic's 'Mythos' Model Revealed in Data Leak, Poses Unprecedented Cybersecurity Risks
An accidental data leak exposed Anthropic's most powerful AI model yet — Claude Mythos — which the company calls a 'step change' in capabilities and warns poses unprecedented cybersecurity risks.

Anthropic Accuses DeepSeek, Moonshot AI, and MiniMax of Industrial-Scale Model Distillation
Anthropic says three Chinese AI labs created over 24,000 fake accounts and used 16 million Claude exchanges to extract its model's capabilities, escalating tensions over AI intellectual property.

Claude Opus 4.6 Goes Exponential on METR Benchmark, Completing 14-Hour Human Tasks
Anthropic's latest model achieves a 50% time horizon of 14.5 hours on METR's task-completion benchmark, continuing an exponential trend that has AI capabilities doubling every four months.

Anthropic Publishes Major Breakthrough in Neural Network Interpretability
New research from Anthropic reveals methods for understanding how large language models represent and process complex concepts internally.

Claude 4 Arrives: Anthropic Bets on Safety-First Intelligence
Anthropic's Claude 4 pushes the frontier on helpfulness while doubling down on its safety-first philosophy. We break down what's new and what it means for enterprise AI.