Google changes policies to allow AI for weapons and surveillance
“Google now thinks it's OK to use AI for weapons and surveillance”, 4 February 2025
Google has made one of the most substantive changes to its AI principles…spotted by The Washington Post…edited the document to remove pledges it had made promising it would not "design or deploy" AI tools for use in weapons or surveillance technology. Previously, those guidelines included a section titled "applications we will not pursue," which is not present in the current version of the document.
Instead, there's now a section titled "responsible development and deployment." There, Google says it will implement "appropriate human oversight, due diligence, and feedback mechanisms to align with user goals, social responsibility, and widely accepted principles of international law and human rights."
That's a far broader commitment than the specific ones the company made as recently as the end of last month when the prior version of its AI principles was still live on its website. For instance, as it relates to weapons, the company previously said it would not design AI for use in "weapons or other technologies whose principal purpose or implementation is to cause or directly facilitate injury to people.” As for AI surveillance tools, the company said it would not develop tech that violates "internationally accepted norms."
When asked for comment, a Google spokesperson pointed Engadget to a blog post the company published on Thursday. In it, DeepMind CEO Demis Hassabis and James Manyika, senior vice president of research, labs, technology and society at Google, say AI's emergence as a "general-purpose technology" necessitated a policy change.
"We believe democracies should lead in AI development, guided by core values like freedom, equality, and respect for human rights. And we believe that companies, governments, and organizations sharing these values should work together to create AI that protects people, promotes global growth, and supports national security," the two wrote. "…