February 25, 2026
Anthropic is narrowing its AI safety policy pledge, removing the company’s previous commitment to halt the development of its AI models if they outpace its safety procedures.
The AI firm unveiled an updated version of its Responsible Scaling Policy on Tuesday, explaining in a blog post that the AI industry has not reached a consensus on risks as it had previously hoped.
“If one AI developer paused development to implement safety measures while others moved forward training and deploying AI systems without strong mitigations, that could result in a world that is less safe—the developers with the weakest protections would set the pace, and responsible developers would lose their ability to do safety research and advance the public benefit,” it said in the updated policy document.
Anthropic said it has opted to separate its own goals from broader industry recommendations on safety going forward.
It also noted that instead of putting forward hard commitments, the company will rely on “nonbinding but publicly-declared” goals that it will grade its progress on.
The update to the Responsible Scaling Policy, which was first released in 2023, comes as Anthropic is currently locked in a dispute with the Pentagon.
At issue is Anthropic’s AI usage policy, which bars the use of its model Claude to conduct mass surveillance or develop weapons that do not require human oversight.
Anthropic CEO Dario Amodei met with Defense Secretary Pete Hegseth at the Pentagon on Tuesday, during which the department threatened to cancel the company’s $200 million contract if it did not agree to the Pentagon’s terms by Friday.
The department also warned that it would use the Defense Production Act against Anthropic or would designate it as a supply chain risk.
No comments:
Post a Comment