Google uses artificial intelligence watermarks to automatically identify text generated by its Gemini chatbot, making it easier to distinguish between AI-generated content and human-written posts. This watermarking system could help prevent AI chatbots from being exploited for misinformation and disinformation, as well as fraud in schools and business environments.
Now, the technology company says it is making available an open-source version of its technology so that other generative AI developers can similarly watermark output from their large-scale language models. I am. Pushmeet Kohli Google DeepMind is the company's AI research team, combining the former Google Brain and DeepMind labs. “SynthID is not a silver bullet for identifying AI-generated content, but it is an important building block for developing more reliable AI identification tools,” he says.
Independent researchers expressed similar optimism. “There is no known way to reliably watermark, but I really think this could help detect some things like AI-generated misinformation and academic fraud,” he said. I say. scott aaronson at the University of Texas at Austin, where he previously worked on AI safety at OpenAI. “We hope that other leading language modeling companies, such as OpenAI and Anthropic, will follow DeepMind’s lead in this regard.”
In May of this year, Google DeepMind announced Google announced that it has implemented the SynthID method for watermarking AI-generated text and video from Google's Gemini and Veo AI services, respectively. The company recently published a paper in the journal nature SynthID generally performs better than similar AI watermarking techniques for text. The comparison involved evaluating how easily the responses from different watermarked AI models were detectable.
In Google DeepMind's AI watermarking approach, as a model generates a sequence of text, a “tournament sampling” algorithm subtly moves it toward selecting “tokens” of specific words that are detectable by associated software. Create a statistical signature. This process randomly pairs candidate word tokens in tournament-style brackets. The winner of each pair is determined by which one gets the highest score according to the watermark function. Winners advance through successive tournament rounds until there is one round remaining. The “layered approach” “further complicates the potential for reverse engineering and attempts to remove watermarks,” it said. Yellow Furong at the University of Maryland.
It said a “determined adversary” with vast computational power could remove such AI watermarks. Hanlin Zhang at Harvard University. But he said SynthID's approach makes sense given the need for scalable watermarking in AI services.
Google DeepMind researchers tested two versions of SynthID that represent a trade-off between making watermark signatures easier to detect in exchange for distorting the text typically produced by AI models. They showed that the undistorted version of the AI watermark continued to work without noticeable impact on the quality of the 20 million text responses Gemini generated during live experiments.
However, the researchers also acknowledged that this watermarking works best on long chatbot responses that can be answered in a variety of ways, such as composing an essay or an email, as well as on math or coding questions. The response to this has not yet been tested.
Google DeepMind's team and others have stated the need for additional safeguards against misuse of AI chatbots, and Huang similarly recommended stronger regulation. “Requiring watermarks by law addresses both practicality and user adoption challenges and makes large language models more secure to use,” she says.
topic:
Source: www.newscientist.com