February 20, 2025
The GIST Editors' notes
This text has been reviewed in accordance with Science X's editorial course of and insurance policies. Editors have highlighted the next attributes whereas making certain the content material's credibility:
fact-checked
peer-reviewed publication
trusted supply
proofread
Assume you may cheat with AI? Researcher creates watermarks to detect AI-generated writing

Synthetic intelligence is placing instructors and employers in a clumsy place in relation to accepting written work, leaving them questioning: Who wrote this? A human or AI?
However think about a digital watermark that might take away the guesswork and truly flag AI-generated textual content each time somebody submits their writing. A College of Florida engineering professor is growing this expertise proper now.
"If I'm a scholar and I'm writing my homework with ChatGPT, I don't need my professor to detect that," stated Yuheng Bu, Ph.D., an assistant professor within the Division of Electrical and Laptop Engineering within the Herbert Wertheim Faculty of Engineering.
Utilizing UF's supercomputer HiPerGator, Bu and his workforce are engaged on an invisible watermark technique for giant language fashions designed to reliably detect AI-generated content material—even altered or paraphrased—whereas sustaining writing high quality.
Navigating the AI panorama
Giant language fashions, akin to Google's Gemini, are AI platforms able to producing human-like textual content. Writers can feed prompts into these AI fashions, and the fashions will full their assignments utilizing info from billions of datasets. This creates a major downside in educational {and professional} settings.
To handle this, Peter Scarfe, Ph.D., and different researchers from the College of Studying in the UK examined AI detection ranges in school rooms final 12 months. They created pretend scholar profiles and wrote their assignments utilizing fundamental AI-generated platforms.
"Total, AI submissions verged on being undetectable, with 94% not being detected," that research famous. "Our 6% detection charge probably overestimates our capability to detect real-world use of AI to cheat on exams."
The low efficiency is as a result of steady development of enormous language fashions, making AI-generated textual content more and more indistinguishable from human-written content material. Consequently, detection turns into progressively harder and will finally grow to be not possible, Bu stated.
Watermarking provides another and efficient answer by proactively embedding particularly designed, invisible alerts into AI-generated textual content. These alerts function verifiable proof of AI era, enabling dependable detection.
Particularly, Bu's work focuses on two key points: sustaining the standard of enormous language model-generated textual content after watermarking, and making certain the watermark's robustness in opposition to varied modifications. The proposed adaptive technique ensures the embedded watermark stays imperceptible to human readers, preserving the pure circulation of writing, in comparison with the unique giant language fashions.
Streamlining the detection course of
Some tech corporations are already growing watermarks for AI-generated textual content. Researchers at Google DeepMind, for instance, created a text-detection watermark final 12 months and deployed it to tens of millions of chatbot customers.
Requested concerning the distinction between these watermarks and his challenge, Bu stated UF's technique "applies watermarks to solely a subset of textual content throughout era, so we imagine it achieves higher textual content high quality and higher robustness in opposition to removing assaults."
Moreover, Bu's work enhances the system's energy in opposition to widespread textual content modifications in each day use, akin to synonym substitute and paraphrasing, which frequently render AI detection instruments ineffective. Even when a person utterly rewrites the watermarked textual content, so long as the semantics stay unchanged, the watermark stays detectable with excessive likelihood. And a watermark secret is utilized by the platform itself.
"The entity that applies the watermark additionally holds the important thing required for detection. If textual content is watermarked by ChatGPT, OpenAI would possess the corresponding key wanted to confirm the watermark," Bu stated. "Finish customers in search of to confirm a watermark should receive the important thing from the watermarking entity. Our strategy employs a personal key mechanism, that means solely the important thing holder can detect and validate the watermark."
The first situation now, Bu stated, is how finish customers receive that watermark key. Within the present framework, a professor should contact the entity that embeds the watermark to acquire the important thing or use an software programming interface supplied by the entity to detect watermarking. The query of who holds the important thing and, consequently, the power to assert mental property, is vital within the improvement of enormous language mannequin watermarking.
"An important subsequent step is to determine a complete ecosystem that enforces watermarking utilization and key distribution or develops extra superior methods that don’t depend on a secret key," Bu stated.
Bu has written a number of papers on AI watermarks, together with "Adaptive Textual content Watermark for Giant Language Fashions" for the Worldwide Convention on Machine Studying (ICML 2024), posted to the arXiv preprint server final 12 months, and "Theoretically Grounded Framework for LLM Watermarking: A Distribution-Adaptive Method," additionally accessible on arXiv.
"Watermarks have the potential to grow to be a vital instrument for belief and authenticity within the period of generative AI," Bu stated. "I see them seamlessly built-in into colleges to confirm educational supplies and throughout digital platforms to tell apart real content material from misinformation. My hope is that widespread adoption will streamline verification and improve confidence within the info we depend on day-after-day."
Extra info: Yepeng Liu et al, Adaptive Textual content Watermark for Giant Language Fashions, arXiv (2024). DOI: 10.48550/arxiv.2401.13927
Haiyun He et al, Theoretically Grounded Framework for LLM Watermarking: A Distribution-Adaptive Method, arXiv (2024). DOI: 10.48550/arxiv.2410.02890
Journal info: PLoS ONE , arXiv Supplied by College of Florida Quotation: Assume you may cheat with AI? Researcher creates watermarks to detect AI-generated writing (2025, February 20) retrieved 20 February 2025 from https://techxplore.com/information/2025-02-ai-watermarks-generated.html This doc is topic to copyright. Aside from any truthful dealing for the aim of personal research or analysis, no half could also be reproduced with out the written permission. The content material is supplied for info functions solely.
Discover additional
Sampling algorithm can 'watermark' AI-generated textual content to indicate the place it got here from 0 shares
Feedback to editors
