Gaming: Anthropic Ditches Its Defining Safety Promise To Pause Dangerous Ai...

Gaming: Anthropic Ditches Its Defining Safety Promise To Pause Dangerous Ai...

There's apparently no point being the only careful AI company.

Keep up to date with the most important stories and the best deals, as picked by the PC Gamer team.

Given the way the AI industry is going these days, the following news probably isn't a huge surprise. But it's unnerving all the same. Announced in a new blog post, Anthropic, arguably the sole remaining example among the major AI players that really bigs up its safety responsibilities, has ditched its core commitment to "pause" development of more powerful AI models if suitable safety safeguards aren’t ready.

In previous versions of what Anthropic calls its Responsible Scaling Policy (RSP), the organisation said that if its AI systems approached certain dangerous capability thresholds—particularly around catastrophic misuse—it would halt further scaling or deployment until adequate safety measures were in place.

To quote Anthropic's original commitment, "the ASL (AI Safety Levels) system implicitly requires us to temporarily pause training of more powerful models if our AI scaling outstrips our ability to comply with the necessary safety procedures."

But that commitment is now gone from Anthropic's newly updated RSP. In Version 3.0 of the RSP, Anthropic has dumped explicit references to “pausing” of development in favour of softer language focused on “responsible development,” “risk management,” and “iterative deployment.”

Instead of promising to stop training models that cross specific danger thresholds, the company now says it will implement safeguards, publish safety evaluations, and release Frontier Safety Framework updates explaining how risks are being handled.

So, why is this happening? Partly, Anthropic seems to be saying, because it's futile being the only AI outfit explicitly committed to safety. "If one AI developer paused development to implement safety measures while others moved forward training and deploying AI systems without strong mitigations, that could result in a world that is less safe—the developers with the weakest protections would set the pace, and responsible developers would lose their ability to do safety research and advance the public benefit," Anthropic's full policy document says.

"We didn't really feel, with the rapid advance of AI, that it made sense for us to make unilateral commitments…if competitors are blazing ahead,” Anthropic’s chief science officer Jared Kaplan also told Time magazine.

Keep up to date with the most important stories and the best

Source: PC Gamer