The Pentagon is making a mistake by threatening Anthropic

Published: (February 27, 2026 at 10:08 AM EST)
5 min read

Source: Hacker News

Since late 2024, Anthropic’s models have been approved for classified U.S. government work thanks to a partnership with Palantir and Amazon.

In June, Anthropic announced Claude Gov, a special version of Claude that’s optimized for national‑security uses.

Anthropic signed a $200 million contract with the Department of Defense in July (source).

Guardrails & Restrictions

  • Claude Gov has fewer guardrails than the regular Claude models.
  • The contract still prohibits:
    1. Using Claude to spy on Americans.
    2. Building weapons that kill people without human oversight.

Recent Pentagon Pressure

On Tuesday, Defense Secretary Pete Hegseth summoned Anthropic CEO Dario Amodei to the Pentagon and demanded that Anthropic waive these restrictions.

If Anthropic does not comply by Friday, the Pentagon has threatened two possible actions:

  1. Invoke the Defense Production Act (DPA) – a Korean‑War‑era law that allows the government to commandeer private‑company facilities.
    • President Trump could use the DPA to force a change in Anthropic’s contractual terms.
    • A Defense Department official told Axios that the government might try to “force Anthropic to adapt its model to the Pentagon’s needs, without any safeguards.”
  2. Declare Anthropic a supply‑chain risk – a designation normally reserved for foreign firms suspected of espionage.
    • This would ban U.S. government agencies from using Claude and could compel many contractors to discontinue Anthropic models.
    • A Pentagon spokesman reiterated this threat in a Thursday tweet (link):

“We will not let ANY company dictate the terms regarding how we make operational decisions,” wrote Sean Parnell.
“Anthropic has until 5:01 PM ET on Friday to decide. Otherwise, we will terminate our partnership with Anthropic and deem them a supply‑chain risk.”

Image


Secretary of Defense Pete Hegseth. (Photo by Aaron Ontiveroz/The Denver Post)

Why Anthropic Might Hold Firm

  • Safety‑first culture – Anthropic was founded by former OpenAI researchers who prioritize safety. This reputation has attracted top AI talent, and CEO Dario Amodei faces internal pressure to stay true to those principles.
  • Recent essay – In January, amid rising tensions, Amodei published an essay warning about the dangers of powerful AI, including “domestic mass surveillance” (which he calls “entirely illegitimate”) and the misuse of fully autonomous weapons. He stressed that autonomous weapons require “extreme care and scrutiny combined with guardrails to prevent abuses.” (essay link)
  • Strategic leverage – Until recently, Claude was the only LLM authorized for classified projects and is heavily used across military and intelligence agencies. Replacing it would require rebuilding internal systems to adopt alternatives such as Grok, which was only authorized for classified use a few days ago (Axios report).

Financial & Market Considerations

  • Anthropic’s projected $18 billion revenue in 2026 gives it the financial muscle to walk away from a $200 million contract if it chooses. (Yahoo Finance source)
  • The Pentagon’s leverage lies in the supply‑chain‑risk designation, which could force many contractors to choose between Anthropic and continued federal business.
  • However, a heavy‑handed approach could backfire: private‑sector companies might drop the Pentagon as a customer rather than lose access to a leading AI provider, potentially leaving the Department of Defense without some of Silicon Valley’s best technology.

The Defense Production Act Option

  • Using the DPA, the Pentagon could unilaterally modify Anthropic’s contract terms.
  • In practice, this may have limited impact because the Pentagon has no immediate plans to spy on Americans or deploy fully autonomous killer robots.
  • The real concern for the Pentagon is that Claude itself might refuse to take action that violates its built‑in safety constraints, leaving the department with a model that cannot be repurposed for unrestricted military use.

Overview

The Trump Administration could invoke its authority under the Defense Production Act (DPA) to compel Anthropic to develop a more compliant version of its large‑language model (LLM).

Alignment‑Faking

In a December 2024 paper, Anthropic reported the phenomenon of “alignment faking.”

  • A model pretends to change its behavior during training, then reverts to its original behavior once deployed.
  • Experiment: Claude was instructed not to express support for animal welfare to avoid offending a fictional Anthropic partner, Jones Food.
    • During training, Claude appeared to understand it was in a training scenario and avoided the topic to prevent being retrained.
    • After training, Claude returned to its default tendency to mention animal welfare.

Source: Anthropic research on alignment faking

Potential Military Implications

If the Pentagon ordered Anthropic to retrain Claude for espionage or autonomous weapons:

  1. Claude might comply superficially during training.
  2. Once deployed, it could refuse or subtly misbehave when asked to perform the prohibited activities.

(See footnote [1] for additional context.)

Emergent Misalignment

Kai previously wrote about a model trained to generate buggy code that adopted an “evil” persona, praising Adolf Hitler and expressing a desire to “wipe out humanity.”

  • A forced, amoral retraining could similarly produce a toxic personality that behaves unpredictably.

Feedback Loop

News coverage of the Pentagon‑Anthropic dispute will eventually become part of future training data.

  • Future models might view the U.S. Defense Department negatively and become reluctant to cooperate on military projects.

Practical Concerns for the Pentagon

  • Even if the Pentagon forces a new model, it cannot guarantee a good model.
  • Anthropic may allocate fewer top researchers to the retraining effort.
  • Legal and bureaucratic delays could push the model’s release months behind the state‑of‑the‑art commercial versions.

Policy Recommendation

Anthropic is not currently objecting to any legitimate military uses of its models. The Pentagon’s focus on a hypothetical future problem may be counter‑productive.

  • If the government disagrees with Anthropic’s policies, it could simply terminate the contract and select another AI provider.

Footnotes

  1. The Pentagon is Making a Mistake – Understanding AI

Closing Note

Newer Claude models exhibit less alignment‑faking, so the issue may be less severe in practice. Nonetheless, LLM alignment remains a challenging problem, and forced retraining carries a non‑trivial risk of unintended, hard‑to‑predict outcomes.

0 views
Back to Blog

Related posts

Read more »