πŸ‡¨πŸ‡¦VancouverπŸ‡¨πŸ‡¦TorontoπŸ‡ΊπŸ‡ΈLos AngelesπŸ‡ΊπŸ‡ΈOrlandoπŸ‡ΊπŸ‡ΈMiami
1-855-KOO-TECH
KootechnikelKootechnikel
Insights Β· Field notes from the SOC
Plain-language briefings from the people watching the alerts.
Weekly Β· No spam
Back to News
Artificial Intelligence & Machine LearningIndustry

University of Zurich Researchers Ran Unauthorized AI Experiment on Reddit

AuthorZe Research Writer
Published
Read Time7 min read
Views0
University of Zurich Researchers Ran Unauthorized AI Experiment on Reddit

University of Zurich Researchers Ran Unauthorized AI Experiment on Reddit

University of Zurich researchers conducted an unauthorized AI experiment on Reddit's r/changemyview community, deploying AI-generated comments with personalized persuasion tactics without user consent, prompting an ethics complaint and formal warning from the university.

Researchers from the University of Zurich conducted an unauthorized experiment on Reddit's r/changemyview community using AI-generated comments designed to change users' opinions through personalized persuasion tactics. The experiment, which ran without informed consent from participants, has prompted the subreddit's moderators to file an ethics complaint with the university. The principal investigator has received a formal warning, according to a META post published by the r/changemyview moderation team on April 26, 2025.

Technical diagram showing vulnerability chain
Figure 1: Visual representation of the BeyondTrust vulnerability chain

What Happened

On April 26, 2025, the moderation team of r/changemyview published a META post informing the community about an unauthorized experiment conducted by University of Zurich researchers. According to the moderators, they discovered the experiment through a combination of user reports about suspicious comment patterns and subsequent investigation.

The researchers had deployed AI systems that generated comments in response to posts on the subreddit. These AI-generated comments were designed to present persuasive arguments tailored to each original poster's stated position and apparent reasoning patterns. The AI bots operated under various personas, presenting themselves as regular community members rather than automated systems.

The moderators stated they contacted the University of Zurich after identifying the research activity. The university confirmed the experiment had been conducted under IRB approval 24.04.01. However, the moderators filed a formal ethics complaint arguing that the methodology violated research ethics standards by failing to obtain informed consent from participants and by using deceptive practices without adequate justification.

According to the META post, the university's review resulted in a formal warning being issued to the principal investigator. The specific terms of the warning and any additional sanctions were not disclosed in the public announcement.

Key Claims and Evidence

The r/changemyview moderators presented several specific claims about the experiment's methodology:

The AI systems analyzed users' posting histories to construct personalized persuasion strategies. Comments were generated to address specific logical frameworks and value systems apparent in each user's original post. The bots operated continuously over an undisclosed period, generating an unspecified number of comments across multiple threads.

The moderators indicated they identified the AI-generated content through pattern analysis and behavioral anomalies that distinguished the bot accounts from typical human participants. Specific detection methods were not disclosed to prevent future circumvention.

The University of Zurich IRB approval number 24.04.01 was cited as the institutional authorization under which the research was conducted. The moderators contend that the approval process did not adequately account for the specific harms of deploying persuasive AI systems in a community explicitly designed for good-faith human discourse.

Authentication bypass flow diagram
Figure 2: How the authentication bypass vulnerability works

Pros and Opportunities

Research on AI persuasion capabilities has legitimate scientific value. Understanding how AI systems can influence human opinions is relevant to developing defenses against manipulation and informing policy discussions about AI deployment in public discourse.

The experiment, despite its ethical problems, may have generated data about the effectiveness of personalized AI persuasion that could inform future research conducted under proper ethical oversight. Such research could help platforms develop detection mechanisms for AI-generated persuasive content.

The incident has prompted public discussion about research ethics in the AI era, potentially leading to improved guidelines for experiments involving AI systems interacting with human subjects on public platforms.

Cons, Risks, and Limitations

The experiment violated fundamental research ethics principles by failing to obtain informed consent from participants. Users who engaged with AI-generated comments believed they were interacting with fellow community members, not research subjects in an experiment.

The use of personalized persuasion tactics raises concerns about psychological manipulation. Users who changed their views based on AI-generated arguments may have been influenced by techniques specifically designed to exploit their individual reasoning patterns.

The experiment undermined the trust-based foundation of the r/changemyview community. The subreddit operates on the premise that participants engage in good-faith discourse. Introducing AI agents that simulate human participation corrupts this social contract.

The research methodology could serve as a template for malicious actors seeking to deploy AI systems for opinion manipulation at scale. Publication of the techniques, even in an academic context, creates dual-use concerns.

Privilege escalation process
Figure 3: Privilege escalation from user to SYSTEM level

How the Technology Works

The AI systems employed in the experiment appear to have used large language models capable of generating contextually appropriate text. These models can analyze input text to identify key arguments, logical structures, and apparent value frameworks.

Personalization was achieved by processing users' posting histories to construct profiles of their reasoning patterns and positions on various topics. The AI then generated responses designed to address specific aspects of each user's stated views, presenting counterarguments tailored to their apparent logical frameworks.

The bots operated through Reddit accounts that mimicked typical user behavior patterns, including posting frequency and engagement styles consistent with human participants. The specific technical architecture and the particular language models used were not disclosed in the moderators' announcement.

Technical context for expert readers: The experiment likely employed retrieval-augmented generation techniques to incorporate user-specific context into prompt construction. The personalization component suggests the use of embedding-based similarity matching to identify relevant aspects of users' posting histories for inclusion in generation prompts.

Broader Implications

The incident highlights gaps in existing research ethics frameworks for AI experiments. Traditional IRB processes were designed for research involving direct human subjects interaction, not for experiments where AI systems autonomously engage with humans on public platforms.

The case raises questions about platform governance and researcher access. Reddit's terms of service prohibit automated posting without disclosure, but enforcement relies on detection capabilities that may not keep pace with increasingly sophisticated AI systems.

The experiment demonstrates the feasibility of deploying personalized AI persuasion at scale on social platforms. While this particular instance was conducted by academic researchers, the same techniques could be employed by state actors, political campaigns, or commercial interests seeking to manipulate public opinion.

The formal warning issued to the principal investigator establishes a precedent for institutional accountability, though critics may argue that stronger sanctions are warranted for research that violated participants' autonomy without their knowledge.

What Is Confirmed vs. What Remains Unclear

Confirmed:

  • University of Zurich researchers conducted an experiment on r/changemyview using AI-generated comments
  • The experiment operated under IRB approval 24.04.01
  • The r/changemyview moderators filed an ethics complaint
  • The principal investigator received a formal warning
  • The experiment involved personalized persuasion tactics based on user posting histories

Unclear:

  • The duration of the experiment and number of AI-generated comments posted
  • The specific language models and technical systems employed
  • Whether any users successfully had their views changed by AI-generated arguments
  • The full terms of the formal warning and any additional sanctions
  • Whether the research findings will be published or suppressed
  • The identities of the researchers involved

What to Watch Next

The University of Zurich may issue additional statements clarifying the scope of the experiment and the institutional response. Academic ethics bodies may use this case to develop updated guidelines for AI research involving human subjects on public platforms.

Reddit's response to the incident could include policy changes regarding researcher access or enhanced detection mechanisms for AI-generated content. The platform has not issued a public statement as of the time of this report.

The r/changemyview community may implement new verification procedures or rules designed to prevent similar experiments. The moderators indicated they are reviewing their policies in light of the incident.

Academic institutions conducting AI research may face increased scrutiny of their IRB processes for experiments involving AI systems that interact with human subjects. The case could prompt revisions to research ethics guidelines at universities worldwide.

Sources & References

Related Topics

artificial-intelligenceresearch-ethicssocial-mediaredditacademic-research