During its inaugural developer conference, Anthropic launched two new AI models the startup claims are among the industry's best, at least in terms of how they score on popular benchmarks.| TechCrunch
Today, we're announcing the Claude 3 model family, which sets new industry benchmarks across a wide range of cognitive tasks. The family includes three state-of-the-art models in ascending order of capability: Claude 3 Haiku, Claude 3 Sonnet, and Claude 3 Opus.| www.anthropic.com
Subscribe • Previous Issues| gradientflow.substack.com
Scaling reinforcement learning, tracing circuits, and the path to fully autonomous agents| www.dwarkesh.com
As AI systems advance in capability, they have the potential to accelerate scientific discovery and drive economic growth. Yet alongside those benefits they also pose a distinct challenge: Highly capable frontier AI systems may introduce or elevate large-scale risks to public safety and national security, including those related to advanced cyber and chemical, biological, radiological, […]| Frontier Model Forum
Today we are publishing a significant update to our Responsible Scaling Policy (RSP), the risk governance framework we use to mitigate potential catastrophic risks from frontier AI systems.| www.anthropic.com
Anthropic is an AI safety and research company that's working to build reliable, interpretable, and steerable AI systems.| www.anthropic.com
DeepSeek, export controls, the future of democracy| www.chinatalk.media
An illustration of a sequence of events where rogue replicating agents emerge and cause harm.| metr.org
By default powerful ML systems will have dangerous capabilities (such as hacking) and may not do what their operators want. Frontier AI labs should design and modify their systems to be less dangerous and more controllable. In particular, labs should:| ailabwatch.org
What labs should do| ailabwatch.org
The three biggest AI companies — Anthropic, OpenAI, and DeepMind — have now all released policies designed to make their AI models less likely to go rogue or cause catastrophic damage as they approach, and eventually exceed, human capabilities. Are they good enough?| 80,000 Hours
Introducing Claude 3.5 Sonnet—our most intelligent model yet. Sonnet now outperforms competitor models and Claude 3 Opus on key evaluations, at twice the speed.| www.anthropic.com
Advanced AI systems could have massive impacts on humanity and potentially pose global catastrophic risks. There are opportunities...| 80,000 Hours