Tech

Five AI labs now let the US government test their models before release. The arrangement is voluntary, has no legal basis, and is the closest thing America has to AI oversight.

In a bid to mitigate AI risks, five leading research labs have begun voluntarily submitting their models to US government testing before public release, effectively establishing a de facto oversight framework in the absence of formal regulations. This unprecedented arrangement, driven by the Mythos crisis, allows the government to assess AI systems' potential national security threats before they reach the public domain. The Commerce Department is now at the forefront of this informal AI governance effort. AI-assisted, human-reviewed.

Overview

Google, Microsoft, and xAI have joined OpenAI and Anthropic in voluntarily giving the US Commerce Department pre-release access to evaluate their AI models. The five companies now account for the vast majority of frontier AI development worldwide, and all five have agreed to let a single government office test their systems before deployment. The arrangement is voluntary, has no statutory basis, and gives the government no power to block a release. It is also the closest thing the United States has to an AI oversight system, and it was built in less than two years by an office with fewer than two hundred staff.

The office

The Center for AI Standards and Innovation sits within the Commerce Department’s National Institute of Standards and Technology. It was established under President Biden in 2023 as the AI Safety Institute, re-established under Trump with a new name and a reorientation toward standards and national security rather than safety research. The centre has completed more than 40 evaluations of AI models, including state-of-the-art systems that have never been released to the public. Developers frequently submit versions with safety guardrails stripped back so that evaluators can probe for national security-relevant capabilities: biological weapon synthesis pathways, cyberattack automation, and autonomous agent behaviours that could be difficult to control at scale.

Chris Fall now directs the centre, following the abrupt departure of Collin Burns, a former AI researcher at Anthropic who was chosen for the role but pushed out by the White House after four days. Burns had left Anthropic, given up valuable stock, and relocated across the country to take the government position. His removal, reportedly driven by his connection to a company the administration was actively fighting, illustrates the political complexity of building an oversight system for an industry where the evaluators and the evaluated come from the same talent pool.

The agreements

The new partnerships with Google, Microsoft, and xAI expand what had been a two-company arrangement into something closer to comprehensive frontier coverage. OpenAI and Anthropic have renegotiated their existing agreements to align with Trump’s AI Action Plan, which directs the centre to lead national security-related model assessments and positions it as part of a broader “evaluations ecosystem.”

The agreements are not contracts. They are voluntary commitments that the companies can withdraw from at any time. No statute requires pre-release evaluation. No regulation gives the centre authority to delay or block deployment. The entire system depends on the AI companies deciding, for their own strategic reasons, that giving the government early access is preferable to the alternative.

The alternative, from the companies’ perspective, is legislation. Several draft bills would give the centre permanent statutory authority, mandatory evaluation requirements, and the power to impose conditions on deployment. The Pentagon has already demonstrated willingness to blacklist AI companies that refuse

Similar Articles

More articles like this

Tech 1 min

OpenAI releases GPT-5.5 Instant, a new default model for ChatGPT

ChatGPT's conversational backbone is undergoing a significant upgrade, as a new, more efficient transformer architecture is set to become the default model for the popular chatbot, leveraging a 55% increase in parameter count over its predecessor to enhance contextual understanding and response accuracy. The GPT-5.5 Instant model is poised to improve the chatbot's ability to engage in nuanced, multi-turn conversations. This shift is expected to elevate the overall user experience. AI-assisted, human-reviewed.

Tech 1 min

OpenAI claims ChatGPT’s new default model hallucinates way less

OpenAI’s GPT-5.5 Instant slashes hallucinations by 52.5% in high-stakes domains—medicine, law, finance—while cutting flagged factual errors by 37.3%, per internal benchmarks. The new default model for ChatGPT now enforces tighter retrieval-augmented grounding and confidence-gated response thresholds, though critics question whether these gains hold under adversarial prompting or real-world deployment. AI-assisted, human-reviewed.

Tech 1 min

Meta sued by major book publishers over copyright infringement

Major book publishers and a prominent author are taking Meta to court over allegations of copyright infringement, claiming the company trained its Llama AI models on pirated copies of their works sourced from notorious websites like LibGen and Sci-Hub, without permission or compensation. The lawsuit accuses Meta of repeatedly copying copyrighted materials, including books and journal articles, to fuel its AI development. The case threatens to upend the use of large language models in content creation. AI-assisted, human-reviewed.

Tech 1 min

Microsoft’s new Xbox shake-up is all about platform changes

Microsoft's Xbox shake-up is driven by a strategic pivot to technical expertise, as new chief Asha Sharma reorganizes the platform team, bringing in veteran Jared Palmer as VP of engineering to spearhead a revamped Xbox engineering strategy, amidst departures and promotions of senior staff. The overhaul aims to bolster the team's technical capabilities, positioning Xbox for future growth and innovation. Key roles are being redefined to focus on AI-driven platform development. AI-assisted, human-reviewed.

Tech 1 min

Half of young Europeans turn to AI to talk about intimate matters

As digital intimacy surges among young Europeans, a growing reliance on AI-powered chatbots for sensitive conversations raises concerns about emotional labor, vulnerability, and the erosion of human connection in the face of increasingly sophisticated language models. The trend highlights a shift in how Europeans navigate intimacy, with AI-facilitated discussions becoming a normalized aspect of relationships. This phenomenon warrants scrutiny of the societal implications. AI-assisted, human-reviewed.

Tech 1 min

From Alan Shepard to Artemis, celebrating 65 years of Americans in space

Sixty-five years after Alan Shepard's pioneering Freedom 7 mission, the US space program has evolved from a single crewed flight to a sustained, high-stakes endeavor with the Artemis program aiming to return humans to the lunar surface by 2025. This milestone marks a significant shift from the early days of space exploration, where a single successful mission could redefine national ambitions. The trajectory of US spaceflight has been shaped by a series of incremental advancements and setbacks. AI-assisted, human-reviewed.