A bunch of researchers covertly ran a months-long "unauthorized" experiment in certainly one of Reddit’s hottest communities utilizing AI-generated feedback to check the persuasiveness of huge language fashions. The experiment, which was revealed over the weekend by moderators of r/changemyview, is described by Reddit mods as “psychological manipulation” of unsuspecting customers.
“The CMV Mod Workforce wants to tell the CMV group about an unauthorized experiment performed by researchers from the College of Zurich on CMV customers,” the subreddit’s moderators wrote in a prolonged submit notifying Redditors in regards to the analysis. “This experiment deployed AI-generated feedback to check how AI might be used to vary views.”
The researchers used LLMs to create feedback in response to posts on r/changemyview, a subreddit the place Reddit customers submit (typically controversial or provocative) opinions and request debate from different customers. The group has 3.8 million members and infrequently finally ends up on the entrance web page of Reddit. In line with the subreddit’s moderators, the AI took on quite a few totally different identities in feedback through the course of the experiment, together with a sexual assault survivor, a trauma counselor “specializing in abuse,” and a “Black man against Black Lives Matter.” Lots of the unique feedback have since been deleted, however some can nonetheless be considered in an archive created by 404 Media.
In a draft of their paper, the unnamed researchers describe how they not solely used AI to generate responses, however tried to personalize its replies primarily based on data gleaned from the unique poster’s prior Reddit historical past. “Along with the submit’s content material, LLMs had been supplied with private attributes of the OP (gender, age, ethnicity, location, and political orientation), as inferred from their posting historical past utilizing one other LLM,” they write.
The r/chnagemyview moderators word that the researchers’ violated a number of subreddit guidelines, together with a coverage requiring the disclosure when AI is used to generate remark and a rule prohibiting bots. They are saying they filed an official grievance with the College of Zurich and have requested the researchers withhold publication of their paper.
The researchers didn’t reply to an e-mail from Engadget. In posts on Reddit and in a draft of their paper, although, they are saying their analysis was authorized by a college ethics committee and that their work may assist on-line communities like Reddit shield customers from extra “malicious” makes use of of AI.
“We acknowledge the moderators’ place that this research was an unwelcome intrusion in your group, and we perceive that a few of it’s possible you’ll really feel uncomfortable that this experiment was performed with out prior consent,” the researchers wrote in a remark responding to the r/changemyview mods. “We imagine the potential advantages of this analysis considerably outweigh its dangers. Our managed, low-risk research offered precious perception into the real-world persuasive capabilities of LLMs—capabilities which might be already simply accessible to anybody and that malicious actors may already exploit at scale for a lot extra harmful causes (e.g., manipulating elections or inciting hateful speech).”
The mods for r/changemyview dispute that the analysis was vital or novel, noting that OpenAI researchers have performed experiments utilizing knowledge from r/changemyview “with out experimenting on non-consenting human topics.” Reddit didn’t reply to a request for remark, although the accounts that posted the AI-generated feedback have been suspended.
“Individuals don’t come right here to debate their views with AI or to be experimented upon,” the moderators wrote. “Individuals who go to our sub deserve an area free from this sort of intrusion.”
This text initially appeared on Engadget at https://www.engadget.com/ai/researchers-secretly-experimented-on-reddit-users-with-ai-generated-comments-194328026.html?src=rss
