Illustrative photo of Pentagon challenging Anthropic's limits on Claude AI for military use during strained contract talks.
Illustrative photo of Pentagon challenging Anthropic's limits on Claude AI for military use during strained contract talks.
AI 生成的图像

Pentagon disputes Anthropic limits on Claude’s military use as contract talks strain

AI 生成的图像
事实核查

After Anthropic CEO Dario Amodei said in late February that the company would not allow its Claude model to be used for mass domestic surveillance or fully autonomous weapons, senior Pentagon officials said they have no intention of using AI for domestic surveillance and insist that private firms cannot set binding limits on how the U.S. military employs AI tools.

Last July, the Pentagon’s chief digital and artificial intelligence officer, Doug Matty, announced contract awards of up to $200 million each to four tech companies—Anthropic, Google, OpenAI, and xAI—to provide advanced AI models for Defense Department missions. Matty said the department intended to speed adoption of commercial AI for “Joint mission essential tasks” in the “warfighting domain,” but the Pentagon released few operational details, citing national security.

The relatively opaque awards drew fresh attention at the end of February, when Anthropic said it was insisting on limits for Claude in a “narrow set of cases.” In a Feb. 26 statement, Amodei said he strongly supported using AI to help defend the United States and other democracies, but argued that some applications could undermine democratic values—including “mass domestic surveillance” and “fully autonomous weapons,” which he described as self-guided combat drones.

Senior Defense Department officials responded by pushing back on both the premise and the company’s leverage. According to reporting cited by The Nation, Pentagon officials said they do not intend to use AI for domestic surveillance and that unmanned weapons systems will remain under human oversight. But they also argued that contractors should not be able to impose their own civil-liberties conditions on Pentagon operations. Emil Michael, the undersecretary of defense for research and engineering, was quoted as saying: “We won’t have any BigTech company decide Americans’ civil liberties.”

The Nation reported that, during negotiations, Michael also raised a separate question about whether Anthropic would oppose the use of Claude in nuclear-related missions such as missile defense, and that Amodei did not object to that use.

The dispute has highlighted a broader tension between the Pentagon’s push to integrate generative AI into intelligence, targeting and weapons development—and the guardrails AI companies say they need to prevent misuse. The Nation pointed to longstanding Defense Department efforts such as Project Maven, which began by using AI to help analyze drone video for potential targets, and DARPA’s Collaborative Operations in Denied Environment (CODE) initiative, which has worked on autonomy for groups of drones operating under preset rules.

Official Pentagon policy on autonomy is outlined in DoD Directive 3000.09, which states that autonomous and semi-autonomous weapons should be designed so commanders and operators can exercise “appropriate levels of human judgment over the use of force.” Critics have argued that the policy’s flexibility still leaves room for autonomy that could significantly reduce real-time human control.

As AI becomes more integrated into military planning and operations, the Anthropic-Pentagon standoff underscores an unresolved question at the center of the U.S. military’s AI expansion: how to reconcile rapid adoption of commercial systems with demands for enforceable limits on domestic surveillance and the delegation of lethal force to machines.

人们在说什么

X discussions reveal a divide on the Pentagon-Anthropic dispute over Claude AI limits. Supporters of Anthropic commend their ethical stance against mass surveillance and autonomous weapons, viewing the Pentagon's blacklisting as overreach. Critics argue private firms cannot impose restrictions on military use and that simple contract terms suffice. Neutral posts detail the standoff, deadlines, and legal escalations, contrasting with OpenAI's compliance. High-engagement accounts from journalists and analysts highlight national security vs. AI safety tensions.

相关文章

Dramatic illustration of Pentagon designating Anthropic's Claude AI a supply chain risk after military usage dispute.
AI 生成的图像

Pentagon designates Anthropic a ‘supply chain risk’ after dispute over military use limits for Claude AI

由 AI 报道 AI 生成的图像 事实核查

The Pentagon has formally notified AI company Anthropic that it is deemed a “supply chain risk,” a rare designation that critics say is typically aimed at adversary-linked technology. The move follows a breakdown in negotiations over whether the U.S. military can use Anthropic’s Claude models for all lawful purposes, versus contractual limits the company says are needed to prevent fully autonomous weapons and mass domestic surveillance.

The Pentagon is considering ending its relationship with AI firm Anthropic due to disagreements over safeguards. Anthropic, the maker of the Claude AI model, has raised concerns about hard limits on fully autonomous weapons and mass domestic surveillance. This stems from the Pentagon's desire to apply AI models in warfighting scenarios, which Anthropic has declined.

由 AI 报道

Anthropic's CEO Dario Amodei stated that the company will not comply with the Pentagon's request to remove safeguards from its AI models, despite threats of exclusion from defense systems. The dispute centers on preventing the AI's use in autonomous weapons and domestic surveillance. The firm, which has a $200 million contract with the Department of Defense, emphasizes its commitment to ethical AI use.

人工智能(AI)已跻身现代战争的核心,在最近的美以对伊朗打击中发挥了作战支持作用。Anthropic 的 Claude 和 Palantir 的 Gotham 被用于情报评估和目标识别。专家预测 AI 在军事应用中的进一步扩展。

由 AI 报道

美国总统唐纳德·特朗普周五表示,他正在指示政府机构停止与Anthropic合作。五角大楼计划将该初创公司宣布为供应链风险,这是在技术护栏争端后的一次重大打击。使用该公司产品的机构将有六个月的逐步淘汰期。

据报道,埃隆·马斯克的 SpaceX 和 xAI 将在一项秘密的五角大楼竞赛中竞争,开发语音控制的自主无人机蜂群技术。该 1 亿美元奖金挑战赛于 1 月启动,将持续六个月。这些公司和五角大楼国防创新单位未回应置评请求。

由 AI 报道

Anthropic has launched a legal plugin for its Claude Cowork tool, prompting concerns among dedicated legal AI providers. The plugin offers useful features for contract review and compliance but falls short of replacing specialized platforms. South African firms face additional hurdles due to data protection regulations.

 

 

 

此网站使用 cookie

我们使用 cookie 进行分析以改进我们的网站。阅读我们的 隐私政策 以获取更多信息。
拒绝