Google removes pledge to not use AI for weapons, surveillance

Google removes pledge to not use AI for weapons, surveillance


Sundar Pichai, CEO of Alphabet Inc., during Stanford’s 2024 Business, Government, and Society forum in Stanford, California, April 3, 2024.

Justin Sullivan | Getty Images

Google has removed a pledge to abstain from using AI for potentially harmful applications, such as weapons and surveillance, according to the company’s updated “AI Principles.”

A prior version of the company’s AI principles said the company would not pursue “weapons or other technologies whose principal purpose or implementation is to cause or directly facilitate injury to people,” and “technologies that gather or use information for surveillance violating internationally accepted norms.”

Those objectives are no longer displayed on its AI Principles website.

“There’s a global competition taking place for AI leadership within an increasingly complex geopolitical landscape,” reads a Tuesday blog post co-written by Demis Hassabis, CEO of Google DeepMind. “We believe democracies should lead in AI development, guided by core values like freedom, equality, and respect for human rights.”

The company’s updated principles reflect Google’s growing ambitions to offer its AI technology and services to more users and clients, which has included governments. The change is also in line with increasing rhetoric out of Silicon Valley leaders about a winner-take-all AI race between the U.S. and China, with Palantir’s CTO Shyam Sankar saying Monday that “it’s going to be a whole-of-nation effort that extends well beyond the DoD in order for us as a nation to win.”

The previous version of the company’s AI principles said Google would “take into account a broad range of social and economic factors.” The new AI principles state Google will “proceed where we believe that the overall likely benefits substantially exceed the foreseeable risks and downsides.”

In its Tuesday blog post, Google said it will “stay consistent with widely accepted principles of international law and human rights — always evaluating specific work by carefully assessing whether the benefits substantially outweigh potential risks.”

The new AI principles were first reported by The Washington Post on Tuesday, ahead of Google’s fourth-quarter earnings. The company’s results missed Wall Street’s revenue expectations and drove shares down as much as 9% in after-hours trading.

Hundreds of protestors including Google workers are gathered in front of Google’s San Francisco offices and shut down traffic at One Market Street block on Thursday evening, demanding an end to its work with the Israeli government, and to protest Israeli attacks on Gaza, in San Francisco, California, United States on December 14, 2023.

Anadolu | Anadolu | Getty Images

Google established its AI principles in 2018 after declining to renew a government contract called Project Maven, which helped the government analyze and interpret drone videos using artificial intelligence. Prior to ending the deal, several thousand employees signed a petition against the contract and dozens resigned in opposition to Google’s involvement. The company also dropped out of the bidding for a $10 billion Pentagon cloud contract in part because the company “couldn’t be sure” it would align with the company’s AI principles, it said at the time.

Touting its AI technology to clients, Pichai’s leadership team has aggressively pursued federal government contracts, which has caused heightened strain in some areas within Google’s outspoken workforce.

“We believe that companies, governments, and organizations sharing these values should work together to create AI that protects people, promotes global growth, and supports national security,” Google’s Tuesday blog post said.

Google last year terminated more than 50 employees after a series of protests against Project Nimbus, a $1.2 billion joint contract with Amazon that provides the Israeli government and military with cloud computing and AI services. Executives repeatedly said the contract didn’t violate any of the company’s “AI principles.”

However, documents and reports showed the company’s agreement allowed for giving Israel AI tools that included image categorization, object tracking, as well as provisions for state-owned weapons manufacturers. The New York Times in December reported that four months prior to signing on to Nimbus, Google officials expressed concern that signing the deal would harm its reputation and that “Google Cloud services could be used for, or linked to, the facilitation of human rights violations.”

Meanwhile, the company had been cracking down on internal discussions around geopolitical conflicts like the war in Gaza.

Google announced updated guidelines for its Memegen internal forum in September that further restricted political discussions about geopolitical content, international relations, military conflicts, economic actions and territorial disputes, according to internal documents viewed by CNBC at the time. 

Google did not immediately respond to a request for comment.

WATCH: Google’s uphill AI battle in 2025

Google's uphill AI battle in 2025



Source

OpenAI in talks to sell around  billion in stock at roughly 0 billion valuation
Technology

OpenAI in talks to sell around $6 billion in stock at roughly $500 billion valuation

Sam Altman, CEO of OpenAI attends the annual Allen and Co. Sun Valley Media and Technology Conference at the Sun Valley Resort in Sun Valley, Idaho, U.S., on July 8, 2025. David A. Grogan | CNBC OpenAI is preparing to sell around $6 billion in stock as part of a secondary sale that would value […]

Read More
Tech IPOs are roaring after ‘years of Prohibition’ — it may be too good
Technology

Tech IPOs are roaring after ‘years of Prohibition’ — it may be too good

Brendan Blumer, Chairman of of Bullish and Tom Farley, CEO of Bullish, Bullish a cryptocurrency exchange operator, pose with staffs during the company’s IPO at the New York Stock Exchange in New York City, U.S., August 13, 2025. NYSE The Bullish IPO this week took on added significance, perhaps because of the company name. When […]

Read More
Sen. Hawley to probe Meta AI bot policies for children following damning report
Technology

Sen. Hawley to probe Meta AI bot policies for children following damning report

Meta Platforms CEO Mark Zuckerberg departs after attending a Federal Trade Commission trial that could force the company to unwind its acquisitions of messaging platform WhatsApp and image-sharing app Instagram, at U.S. District Court in Washington, D.C., U.S., April 15, 2025. Nathan Howard | Reuters Sen. Josh Hawley, R-Mo., said Friday that he will investigate […]

Read More