In September, Anthropic published their Responsible Scaling Policy outlining the protocols they will be adopting to help manage “increasingly capable AI Systems.”
There are rumours and whispers about OpenAI supposedly having a secret AGI they’ve been hiding from the public; a salient take on responsible policies to prevent such a thing from happening seems prudent.
We’ll take a critical look at the AI Safety Levels (modeled after the US government’s biosafety level standards), discuss whether these safety standards are enough, and explore the possible industry impacts as other companies are expected to fall in line.
One challenge with the ASL scheme as compared to BSL is that ASLs above our current capabilities represent systems that have never been built before – in contrast to BSL, where the highest levels include specific dangerous pathogens that exist today. The ASL system thus has an unavoidable component of “building the airplane while flying it”— we will have to start acting on many provisions of this policy before others can reasonably be specified.