OpenAI, Anthropic agree to have their models tested before making them public

Not so fast, says the U.S. government.
 By 
Cecily Mauran
 on 
ChatGPT OpenAI website through magnifying glass on screen
OpenAI and Anthropic are now subject to new levels of scrutiny. Credit: Rokas Tenys / Shutterstock

OpenAI and rival company Anthropic have signed agreements with the U.S. government to have new models tested before public release.

On Thursday the National Institute of Standards and Technology (NIST) announced that its AI Safety Institute will oversee "AI safety research, testing and evaluation" with both companies. "These agreements are just the start, but they are an important milestone as we work to help responsibly steward the future of AI," said Elizabeth Kelly, director of the AI Safety Institute in the announcement.

It's no secret that generative AI poses safety risks. Its tendency to produce inaccuracies and misinformation, enable harmful or illegal behavior, and entrench discrimination and biases is well documented at this point. OpenAI has its own internal safety testing, but has been secretive about how its models work and what they're trained on. This is the first instance of OpenAI opening up access to third party scrutiny and accountability. Altman and OpenAI have been vocal about the need for AI regulation and standardization. But critics say the willingness to work with the government is a strategy to ensure OpenAI is regulated favorably and stamps out competition.


You May Also Like

"For many reasons, we think it's important that this happens at the national level. US needs to continue to lead!" posted OpenAI CEO Sam Altman on X.

The formal collaboration with NIST builds on the Biden Administrations AI executive order that was signed last October. Amongst other mandates that tapped several federal agencies to ensure the safe and responsible deployment of AI, the order directed requires AI companies to grant access to NIST for red-teaming before an AI model is released to the public.

The announcement also said that it would share findings and feedback in partnership with the UK AI Safety Institute.

Mashable Image
Cecily Mauran
Tech Reporter

Cecily is a tech reporter at Mashable who covers AI, Apple, and emerging tech trends. Before getting her master's degree at Columbia Journalism School, she spent several years working with startups and social impact businesses for Unreasonable Group and B Lab. Before that, she co-founded a startup consulting business for emerging entrepreneurial hubs in South America, Europe, and Asia. You can find her on X at @cecily_mauran.

Mashable Potato

Recommended For You
Not so fast: Anthropic and US military might do business after all
Anthropic logo

Anthropic: Chinese AI firms created 24,000 fraudulent accounts for 'distillation attacks'
Deepseek logo is displayed on a mobile phone screen with the flag of China in background

Anthropic challenges Department of War designation as AI dispute escalates
Anthropic logo on mobile device

Meet Claude Mythos: Leaked Anthropic post reveals the powerful upcoming model
Claude by Anthropic on smartphone

Anthropic Super Bowl LX ads mock ChatGPT
screenshot of anthropic super bowl lx ads featuring handsome black actor and words 'ads are coming to chatgpt. but not to claude.'

Trending on Mashable
NYT Connections hints today: Clues, answers for April 3, 2026
Connections game on a smartphone

Wordle today: Answer, hints for April 3, 2026
Wordle game on a smartphone


What's new to streaming this week? (April 3, 2026)
A composite of images from film and TV streaming this week.

NYT Connections hints today: Clues, answers for April 2, 2026
Connections game on a smartphone
The biggest stories of the day delivered to your inbox.
These newsletters may contain advertising, deals, or affiliate links. By clicking Subscribe, you confirm you are 16+ and agree to our Terms of Use and Privacy Policy.
Thanks for signing up. See you at your inbox!