The probability that one word will follow another can be used to create a watermark for AI-generated text Vikram Arun/Shutterstock
Google has been using artificial intelligence watermarking to automatically identify text generated by the company鈥檚 Gemini chatbot, making it easier to distinguish AI-generated content from human-written posts. That watermark system could help prevent misuse of the AI chatbots for misinformation and disinformation 鈥 not to mention cheating in school and business settings.
Now, the tech company is making an open-source version of its technique available so that other generative AI developers can similarly watermark the output from their own large language models, says at Google DeepMind, the company鈥檚 AI research team, which combines the former Google Brain and DeepMind labs. 鈥淲hile SynthID isn鈥檛 a silver bullet for identifying AI-generated content, it is an important building block for developing more reliable AI identification tools,鈥 he says.
Advertisement
Independent researchers voiced similar optimism. 鈥淲hile no known watermarking method is foolproof, I really think this can help in catching some fraction of AI-generated misinformation, academic cheating and more,鈥 says at The University of Texas at Austin, who previously worked on AI safety at OpenAI. 鈥淚 hope that other large language model companies, including OpenAI and Anthropic, will follow DeepMind’s lead on this.鈥
In May of this year, Google DeepMind that it had implemented its SynthID method for watermarking AI-generated text and video from Google鈥檚 Gemini and Veo AI services, respectively. The company has now published a paper in the journal Nature聽showing how SynthID generally outperformed similar AI watermarking techniques for text. The comparison involved assessing how readily responses from various watermarked AI models could be detected.
In Google DeepMind鈥檚 AI watermarking approach, as the model generates a sequence of text, a 鈥渢ournament sampling鈥 algorithm subtly nudges it toward selecting certain word 鈥渢okens鈥, creating a statistical signature that is detectable by associated software. This process randomly pairs up possible word tokens in a tournament-style bracket, with the winner of each pair being determined by which one scores highest according to a watermarking function. The winners move through successive tournament rounds until just one remains 鈥 a 鈥渕ulti-layered approach鈥 that 鈥渋ncreases the complexity of any potential attempts to reverse-engineer or remove the watermark鈥, says at the University of Maryland.
Free newsletter
Sign up to The Weekly
The best of New 女生小视频, including long-reads, culture, podcasts and news, each week.

A 鈥渄etermined adversary鈥 with huge amounts of computational power could still remove such AI watermarks, says at Harvard University. But he described SynthID鈥檚 approach as making sense given the need for scalable watermarking in AI services.
The Google DeepMind researchers tested two versions of SynthID that represent trade-offs between making the watermark signature more detectable, at the expense of distorting the text typically generated by an AI model. They showed that the non-distortionary version of the AI watermark still worked, without noticeably affecting the quality of 20 million Gemini-generated text responses during a live experiment.
But the researchers also acknowledged that the watermarking works best with longer chatbot responses that can be answered in a variety of ways 鈥 such as generating an essay or email 鈥 and said it has not yet been tested on responses to maths or coding problems.
Both Google DeepMind鈥檚 team and others described the need for additional safeguards against misuse of AI chatbots 鈥 with Huang recommending stronger regulation as well. 鈥淢andating watermarking by law would address both the practicality and user adoption challenges, ensuring a more secure use of large language models,鈥 she says.
Journal reference
Nature
Topics:



