OpenAI and Anthropic, two leading AI startups, have agreed to allow the U.S. AI Safety Institute to test their models before public release. This move follows growing industry concerns about AI safety and ethics. The institute, part of the National Institute of Standards and Technology, will gain early access to these models to assess capabilities and risks.
“We are happy to have reached an agreement with the US AI Safety Institute for pre-release testing of our future models,” OpenAI CEO Sam Altman wrote in a post on X.
While Jack Clark, co-founder of Anthropic, said the company’s “collaboration with the U.S. AI Safety Institute leverages their wide expertise to rigorously test our models before widespread deployment” and “strengthens our ability to identify and mitigate risks, advancing responsible AI development.”
This partnership comes as OpenAI pursues a new funding round, potentially valuing the company at over $100 billion. Meanwhile, Anthropic, valued at $18.4 billion, continues to attract significant investment, including from Amazon.
The agreement aims to foster collaborative research to improve AI safety measures as the industry rapidly evolves.