A Pivotal Day for AI Safety

In what critics are calling a watershed moment for the artificial intelligence industry, Anthropic announced sweeping changes to its Responsible Scaling Policy on Tuesday, eliminating the hard safety tripwires that had been central to the company's identity since its founding. The timing was striking — the announcement arrived on the very same day that reports surfaced about Defense Secretary Pete Hegseth pressuring the company to give the U.S. military unfettered access to its Claude AI model.

For more than two years, Anthropic's RSP stood as one of the most concrete safety commitments in the AI industry. The policy established clear red lines: if the company's models reached certain capability thresholds without adequate safety measures in place, development would stop. That pledge is now gone, replaced by a more flexible framework of "Risk Reports" and "Frontier Safety Roadmaps" that the company says better reflects the realities of the competitive AI landscape.

The Rationale Behind the Shift

Anthropic framed the changes as a pragmatic response to a collective action problem. "Two and a half years later, our honest assessment is that some parts of this theory of change have played out as we hoped, but others have not," the company wrote in its updated policy document. The core argument is straightforward: if one responsible developer pauses while competitors race ahead, the result could be a world shaped by the least careful actors rather than the most thoughtful ones.

"We felt that it wouldn't actually help anyone for us to stop training AI models," Jared Kaplan, Anthropic's chief science officer, told Time magazine. "We didn't really feel, with the rapid advance of AI, that it made sense for us to make unilateral commitments… if competitors are blazing ahead." It's a familiar argument in technology — the idea that responsible actors need to remain at the frontier to ensure safety-minded perspectives shape how powerful technology develops.

But the reasoning sits uneasily alongside the company's rising commercial fortunes. Anthropic raised $30 billion in new investment just this month, bringing its valuation to $380 billion. Its Claude models have drawn widespread acclaim, particularly for coding applications. The latest versions have been described by the company itself as its safest yet — raising the question of why safety pledges need weakening at precisely the moment when capabilities and resources are at their peak.