In a strategically significant legal move, Microsoft has filed a court brief supporting Anthropic in its challenge to the Pentagon’s supply-chain risk designation, directly challenging the government’s claimed right to define the ethical limits of AI use in military contexts. The brief was submitted to a federal court in San Francisco and called for a temporary restraining order. Amazon, Google, Apple, and OpenAI have also backed Anthropic through a separate court filing.
The dispute originated in a $200 million contract negotiation that fell apart after Anthropic refused to allow its Claude AI to be used for mass surveillance or autonomous lethal weapons. Defense Secretary Pete Hegseth labeled the company a supply-chain risk after talks collapsed, and Anthropic’s existing government contracts began to be cancelled. The company filed two simultaneous lawsuits in California and Washington DC, arguing the designation was both unconstitutional and unprecedented for a US firm.
Microsoft’s strategic intervention is informed by its direct integration of Anthropic’s AI into military systems and its participation in the Pentagon’s $9 billion cloud computing contract. The company also holds additional federal agreements with defense, intelligence, and civilian agencies. Microsoft publicly argued that the government and technology sector needed to work together to ensure advanced AI served national security responsibly.
Anthropic’s court filings argued that the supply-chain risk designation was an unconstitutional act of retaliation for the company’s public advocacy of responsible AI development. The company disclosed that it does not currently believe Claude is safe or reliable enough for lethal autonomous operations, which it said was the genuine basis for its contract demands. The Pentagon’s technology chief publicly foreclosed any possibility of renewed negotiations.
Congressional Democrats have separately asked the Pentagon whether AI was involved in a strike in Iran that reportedly killed over 175 civilians at a school, raising questions about AI targeting systems and human oversight. Their formal inquiries are adding legislative pressure to an already extraordinary legal confrontation. Together, these developments are forcing a fundamental reckoning with the question of who has the authority to set the ethical limits on AI in American warfare.
Microsoft’s Strategic Court Move for Anthropic Challenges the Pentagon’s Right to Define AI Ethics
1