Google has finalized a classified AI agreement with the Pentagon for
**TL;DR** Google entered into a classified AI agreement with the Pentagon for “any lawful government purpose” just one day after over 580 employees urged CEO Sundar Pichai to decline such an arrangement. The contract includes advisory restrictions (no mass surveillance and no autonomous weapons without human oversight), but allows the government to request changes to safety settings. On the same day, Bloomberg reported that Google quietly exited a $100 million contest for drone swarm technology in February after an internal ethics review. Google is attempting to differentiate between providing general-purpose AI and creating specific weapons, though this distinction may not hold up on classified networks.
Google has confirmed a deal that allows the Pentagon to utilize its Gemini AI models for classified military operations under terms that permit "any lawful government purpose." This announcement came a day after more than 580 Google employees signed a letter urging Pichai to reject such agreements. The contract grants the Department of Defense API access to Google's AI systems on classified networks, building on a relationship that already deploys Gemini technology to three million Pentagon personnel on unclassified systems. The agreement includes a clause stating that “the AI System is not intended for, and should not be used for, domestic mass surveillance or autonomous weapons (including target selection) without appropriate human oversight and control.” Concurrently, Bloomberg disclosed that Google had discreetly withdrawn from a $100 million Pentagon prize challenge aimed at developing technology for voice-controlled autonomous drone swarms, retracting in February after an internal ethics review despite being advanced in the competition. Google officially cited a lack of “resourcing.” The company attempts to draw a line, but it is not the line that the employees requested.
The classified AI agreement functions as an extension of an existing Pentagon contract, providing API access instead of custom model development or specialized military applications. A representative from Google Public Sector confirmed this arrangement. The Pentagon will have direct access to Google’s software on classified networks that are separated from the public internet and used for mission planning, intelligence analysis, and weapons targeting. The language permitting “any lawful government purpose” aligns Google with OpenAI and Elon Musk’s xAI, both of which have formed their own classified AI agreements with the Pentagon. The government can request modifications to Google’s AI safety protocols and content filters, effectively allowing the Pentagon to alter the safeguards established by Google’s researchers.
The stated restrictions against mass surveillance and autonomous weapons without human oversight reflect similar conditions negotiated by OpenAI in its Pentagon deal. However, the enforcement mechanism remains the same one highlighted by Google's employees as inadequate in their letter: on classified networks, Google is unable to monitor the queries executed, the outputs generated, or the decisions made based on those outputs. The phrase "should not be used for" is advisory rather than a binding prohibition, and “appropriate human oversight and control” lacks clear definition. The employees stated that “the only way to guarantee that Google does not become associated with such harms is to reject any classified workloads.” Google opted to accept these workloads using language that employees had previously argued was unenforceable. Pichai opened Cloud Next 2026 announcing 750 million Gemini users and a $240 billion backlog. This same Gemini infrastructure catering to those users is now being made available on classified military networks, where its usage cannot be monitored by anyone outside the Pentagon.
The exit from the drone swarm challenge is another aspect of this story. Google progressed in a $100 million Pentagon prize challenge to develop technology that would enable commanders to manage autonomous drone swarms using voice commands, converting spoken instructions like “left” into digital signals sent to the drones. The company officially notified the government on February 11, 2026, of its decision to withdraw. While Google cited a lack of resources, Bloomberg's findings indicate that this choice followed an internal ethics review. This withdrawal is reminiscent of Project Maven in 2018, when around 4,000 Google employees protested against the use of AI for analyzing drone video feeds, leading Google to allow the contract to expire. Palantir took over the contract, which was valued at millions of dollars, with Palantir’s stake in the Maven project subsequently growing to $13 billion.
The contrast is striking. On the same day that Google confirmed a deal granting the Pentagon classified access to Gemini for “any lawful government purpose,” it also announced its departure from a program that would have applied its AI to control autonomous drone swarms. Google is prepared to place its most powerful AI models on classified networks without oversight, yet it hesitated to develop voice-controlled drone swarms. This distinction is critical to Google’s internal ethical guidelines: API access to general-purpose models is a step removed from weapons applications, even if the models will be utilized on networks that engage in weapons targeting. In contrast, creating technology expressly designed for managing drone swarms constitutes a direct weapons application that the ethics review could not sanction. The line Google seeks to draw is between supplying tools and creating weapons, between providing access and designing lethal technology. The question hinges on
Other articles
Google has finalized a classified AI agreement with the Pentagon for
Google provided the Pentagon with classified access to Gemini for "any lawful purpose" just one day after more than 580 employees voiced their objections. Additionally, the company withdrew from a $100 million drone swarm competition following an ethics assessment.
