AI Labs Should Pass Safety Review to Get US Government Contracts, Group Says

May 11 (Reuters) – The ⁠Trump ⁠administration should screen cutting-edge ⁠artificial intelligence models for security threats before ​they are publicly released and withhold lucrative government contracts ‌from those that fail ‌review, an advocacy group told U.S. officials on ⁠Monday.

The ⁠White House is grappling with the implications of Anthropic’s ​Mythos, which could make complex cyberattacks easier and quicker to execute, posing national security risks.

Americans for Responsible Innovation urged ​the Trump administration to develop methods to vet upcoming ⁠frontier models ⁠from larger developers ⁠for ​cyberattack and weapons development capabilities.

Companies should have to pass the ​review to be ⁠eligible for government contracts, the group said in a letter to administration officials.

The U.S. Center for AI Standards and Innovation already reviews some AI models through ⁠voluntary agreements with OpenAI, Anthropic, and, more recently, Google, Microsoft ⁠and xAI.

CAISI should take the lead on developing mandatory requirements, and Congress should create a permanent enforcement office within the U.S. Department of Commerce to enforce the requirements, the group said.

The proposed requirements would apply to companies that spend $100 million or more a year on compute to train ⁠frontier models, or that make at least $500 million in revenue annually from AI products and services.

California has a similar threshold for safety reporting requirements ​enacted last year.

(Reporting by Jody Godoy in ​New YorkEditing by Rod Nickel)

Copyright 2026 Thomson Reuters.

Leave a Comment