The sudden wind-down of Anthropic technology inside the U.S. authorities is elevating issues about whether or not federal officers, with out entry to Claude, would possibly fall behind within the quest to protect towards the specter of AI-generated or AI-assisted nuclear and chemical weapons.
Although the rollout has been messy—and Claude stays in use in some components of the federal government—the Trump administration’s anti-Anthropic posture may have a chilling impact on collaborations between AI firms and federal businesses, together with partnerships targeted on vital nationwide safety questions associated to those sorts of futuristic threats, a number of sources inform Quick Firm. The concern is that severing ties with the corporate may each restrict authorities researchers’ understanding of how, sooner or later, unhealthy actors may use AI to generate new forms of nuclear and organic weapons, and maintain again scientific progress extra broadly.
Since a minimum of February 2024, Anthropic has participated in a proper partnership with the National Nuclear Security Administration, the federal company charged with monitoring the nation’s nuclear stockpile. The purpose of that work, the corporate has beforehand mentioned, is to “consider our AI fashions for potential nuclear and radiological dangers.” The priority, right here, is that growing nuclear weapons requires specialised information, however that AI, because it continues to advance, may finally turn into adept at growing this experience by itself. Ultimately, a big language mannequin would possibly be capable to assist somebody determine the right way to design an extremely harmful weapon—and even provide you with a novel one itself.
Now, within the wake of President Donald Trump’s Reality Social publish demanding that federal staff cease utilizing Anthropic tech, it’s not clear what would possibly occur to Anthropic’s efforts to protect towards these future threats. Some federal businesses seem to nonetheless be weighing the right way to strategy the Claude use circumstances they have already got, whereas others are slicing off entry to the software totally.
“As directed by President Trump, the Division of Power is reviewing all present contracts and makes use of of Anthropic expertise,” a spokesperson for the NNSA tells Quick Firm. “The Division stays firmly dedicated to making sure that the expertise we make use of serves the general public curiosity, protects America’s power and nationwide safety, and advances our mission.” Anthropic declined to remark.
Security issues on the Power Division
For the previous few years, Anthropic has been collaborating with or offering expertise to the myriad businesses and nationwide labs that fall beneath the Division of Power. As an illustration, the Lawrence Livermore Nationwide Laboratory started using Claude for Enterprise in 2025 and, on the time, made the software obtainable to about 10,000 scientists. The lab mentioned final yr that the expertise was supposed to assist speed up its analysis efforts “within the domains of nuclear deterrence, power safety, supplies science,” and different areas.
Anthropic has additionally labored with the Nationwide Nuclear Safety Company on evaluating potential AI-related nuclear security dangers. For instance, the company has offered Anthropic with “high-level” metrics and steering which have helped the corporate analyze the specter of its personal expertise. Anthropic has additionally labored with the NNSA on developing technology that may scan and categorize AI chatbot conversations and seek for indicators that somebody may be utilizing an LLM to debate constructing a nuclear weapon.
A listing for 2025 for the Division of Power disclosed that the company was utilizing Claude on the Pacific Northwest Nationwide Laboratory, the Lawrence Livermore Nationwide Laboratory, and the Idaho Nationwide Laboratory in pilots. Anthropic can be certainly one of a number of companions within the company’s Genesis mission, which goals to speed up scientific improvement by leveraging synthetic intelligence.
These collaborations might now be in jeopardy. Claude is “in all places” within the Power Division’s labs, together with on the NNSA, in keeping with Ann Dunkin, the division’s former chief data officer. If labs, or the NNSA, “are engaged on tasks utilizing Anthropic as their AI software, they’re going to must, on the very least, cease work and begin with a brand new vendor,” Durking tells Quick Firm. “It will price money and time. Greater than probably, there might be [new] work as they should practice a brand new mannequin.” To conduct simulations that contain learning varied AI dangers, it’s vital to know how all AI fashions would possibly behave, she provides.
In regard to nuclear weapons, there’s fear that AI might be used to assemble sufficient data to construct one such weapon—or be jailbroken in order that it may present that data, Dunkin says.
A former Division of Homeland Safety official who targeted on AI issues of safety echoes these issues. Anthropic, the particular person tells Quick Firm, was a pacesetter on evaluating how AI fashions, together with its personal, would possibly create severe security dangers associated to chemical and nuclear weapons. Stress to take away Anthropic dangers losing peoples’ time and is probably not profitable anyway, they are saying. It additionally places federal officers behind on attempting to know the total dangers associated to synthetic intelligence, or to totally profit from its efficiencies, provided that Anthropic continues to be the main supplier of some AI capabilities. “There’s no ban on Claude for the unhealthy guys,” the previous official provides.
General, the federal government’s sudden flip towards Anthropic dangers scaring off different firms that may need to work on severe points, together with these associated to nuclear safety. “Anthropic discovered that after you’re serving the U.S. authorities, you won’t have the appropriate to say no, a minimum of now with out retaliation. Naturally that may deter others from working for the federal government, particularly on delicate matters,” says Steven Adler, an ex-OpenAI worker who focuses on AI issues of safety.
“There’s a bitter irony right here: The administration is concurrently demanding AI firms assist with nationwide safety and making it more durable for accountable actors to do precisely that,” Alex Bores, who’s operating for a Home seat in New York on a platform targeted on AI regulation, tells Quick Firm in an announcement. “AI firms working with NNSA to guage danger isn’t a legal responsibility—it’s a mannequin. Punishing it sends precisely the improper sign at precisely the improper time.”
An incomplete exit plan
It’s not instantly clear how federal businesses are purported to strategy Anthropic expertise proper now. Trump used Reality Social to demand that federal businesses “instantly stop all use of Anthropic’s expertise,” although such directions are ordinarily communicated by the federal chief data officer. The Trump administration is reportedly engaged on an govt order associated to Anthropic, whereas Anthropic has filed a lawsuit difficult its designation as a “provide chain danger.”
The Normal Companies Administration, in keeping with one post, appears to be deciphering the Reality Social publish as a nationwide safety directive. The company’s GitHub repository reveals that Claude was not too long ago eliminated for its interagency AI useful resource, and an individual inside the company confirmed that staff may not entry Claude internally. Nonetheless, one other particular person on the company tells Quick Firm that no official directions about the right way to really implement eradicating Claude from federal use circumstances have really been despatched to staff.
One main problem with stripping Anthropic’s expertise from the federal authorities is that the expertise might be delivered in some ways. In Claude’s case, this contains merchandise bought by Anthropic instantly, but additionally integrations with widespread—and controversial—authorities contractors like Palantir and Amazon Internet Companies.
Notably, Claude for Government continues to be listed as one of many options provided inside the Palantir Federal Cloud Service, and a number of other businesses have licensed the usage of this package deal, together with the Brookhaven Nationwide Lab and the Environmental Administration Consolidated Enterprise Middle, in addition to the State Division and the Treasury. The product describes Claude as “objective constructed” to fulfill excessive authorities safety necessities. Palantir additionally has a long-standing relationship with the NNSA that predates LLMs.
The NNSA spokesperson declined to touch upon how they had been approaching the usage of Claude and categorized techniques. On the time of this writing, Palantir had not respondd to a request for remark.
On the army aspect of presidency, a lot ado has been product of the truth that solely Caude, and never techniques like ChatGPT, has been cleared to function in categorized environments. The Pentagon has since despatched a memo to staff that prioritizes removing Claude from any techniques that contain nuclear safety. Labeled environments are additionally vital to civilian businesses. Although Treasury Secretary Scott Bessent has mentioned his company might be “terminating” use of Anthropic merchandise and Claude, there was a minimum of some grumbling at a latest assembly targeted on AI use inside the company that different AI instruments weren’t equally obtainable for categorized data.

