Back to All Events

Anthropic’s Responsible Scaling Policy

In September, Anthropic published their Responsible Scaling Policy outlining the protocols they will be adopting to help manage “increasingly capable AI Systems.”

There are rumours and whispers about OpenAI supposedly having a secret AGI they’ve been hiding from the public; a salient take on responsible policies to prevent such a thing from happening seems prudent.

We’ll take a critical look at the AI Safety Levels (modeled after the US government’s biosafety level standards), discuss whether these safety standards are enough, and explore the possible industry impacts as other companies are expected to fall in line.

One challenge with the ASL scheme as compared to BSL is that ASLs above our current capabilities represent systems that have never been built before – in contrast to BSL, where the highest levels include specific dangerous pathogens that exist today. The ASL system thus has an unavoidable component of “building the airplane while flying it”— we will have to start acting on many provisions of this policy before others can reasonably be specified.

Anthropic’s Responsible Scaling Policy

Previous
Previous
29 November

The OpenAI Debacle: what does this mean for the future of AGI development?

Next
Next
13 December

Progress Measures for Grokking via Mechanistic Interpretability