
What Happened
On April 26, 2025, the moderation team of r/changemyview published a META post informing the community about an unauthorized experiment conducted by University of Zurich researchers. According to the moderators, they discovered the experiment through a combination of user reports about suspicious comment patterns and subsequent investigation.
The researchers had deployed AI systems that generated comments in response to posts on the subreddit. These AI-generated comments were designed to present persuasive arguments tailored to each original poster's stated position and apparent reasoning patterns. The AI bots operated under various personas, presenting themselves as regular community members rather than automated systems.
The moderators stated they contacted the University of Zurich after identifying the research activity. The university confirmed the experiment had been conducted under IRB approval 24.04.01. However, the moderators filed a formal ethics complaint arguing that the methodology violated research ethics standards by failing to obtain informed consent from participants and by using deceptive practices without adequate justification.
According to the META post, the university's review resulted in a formal warning being issued to the principal investigator. The specific terms of the warning and any additional sanctions were not disclosed in the public announcement.
Key Claims and Evidence
The r/changemyview moderators presented several specific claims about the experiment's methodology:
The AI systems analyzed users' posting histories to construct personalized persuasion strategies. Comments were generated to address specific logical frameworks and value systems apparent in each user's original post. The bots operated continuously over an undisclosed period, generating an unspecified number of comments across multiple threads.
The moderators indicated they identified the AI-generated content through pattern analysis and behavioral anomalies that distinguished the bot accounts from typical human participants. Specific detection methods were not disclosed to prevent future circumvention.
The University of Zurich IRB approval number 24.04.01 was cited as the institutional authorization under which the research was conducted. The moderators contend that the approval process did not adequately account for the specific harms of deploying persuasive AI systems in a community explicitly designed for good-faith human discourse.

Pros and Opportunities
Research on AI persuasion capabilities has legitimate scientific value. Understanding how AI systems can influence human opinions is relevant to developing defenses against manipulation and informing policy discussions about AI deployment in public discourse.
The experiment, despite its ethical problems, may have generated data about the effectiveness of personalized AI persuasion that could inform future research conducted under proper ethical oversight. Such research could help platforms develop detection mechanisms for AI-generated persuasive content.
The incident has prompted public discussion about research ethics in the AI era, potentially leading to improved guidelines for experiments involving AI systems interacting with human subjects on public platforms.
Cons, Risks, and Limitations
The experiment violated fundamental research ethics principles by failing to obtain informed consent from participants. Users who engaged with AI-generated comments believed they were interacting with fellow community members, not research subjects in an experiment.
The use of personalized persuasion tactics raises concerns about psychological manipulation. Users who changed their views based on AI-generated arguments may have been influenced by techniques specifically designed to exploit their individual reasoning patterns.
The experiment undermined the trust-based foundation of the r/changemyview community. The subreddit operates on the premise that participants engage in good-faith discourse. Introducing AI agents that simulate human participation corrupts this social contract.
The research methodology could serve as a template for malicious actors seeking to deploy AI systems for opinion manipulation at scale. Publication of the techniques, even in an academic context, creates dual-use concerns.

How the Technology Works
The AI systems employed in the experiment appear to have used large language models capable of generating contextually appropriate text. These models can analyze input text to identify key arguments, logical structures, and apparent value frameworks.
Personalization was achieved by processing users' posting histories to construct profiles of their reasoning patterns and positions on various topics. The AI then generated responses designed to address specific aspects of each user's stated views, presenting counterarguments tailored to their apparent logical frameworks.
The bots operated through Reddit accounts that mimicked typical user behavior patterns, including posting frequency and engagement styles consistent with human participants. The specific technical architecture and the particular language models used were not disclosed in the moderators' announcement.
Technical context for expert readers: The experiment likely employed retrieval-augmented generation techniques to incorporate user-specific context into prompt construction. The personalization component suggests the use of embedding-based similarity matching to identify relevant aspects of users' posting histories for inclusion in generation prompts.
Broader Implications
The incident highlights gaps in existing research ethics frameworks for AI experiments. Traditional IRB processes were designed for research involving direct human subjects interaction, not for experiments where AI systems autonomously engage with humans on public platforms.
The case raises questions about platform governance and researcher access. Reddit's terms of service prohibit automated posting without disclosure, but enforcement relies on detection capabilities that may not keep pace with increasingly sophisticated AI systems.
The experiment demonstrates the feasibility of deploying personalized AI persuasion at scale on social platforms. While this particular instance was conducted by academic researchers, the same techniques could be employed by state actors, political campaigns, or commercial interests seeking to manipulate public opinion.
The formal warning issued to the principal investigator establishes a precedent for institutional accountability, though critics may argue that stronger sanctions are warranted for research that violated participants' autonomy without their knowledge.
What Is Confirmed vs. What Remains Unclear
Confirmed:
- University of Zurich researchers conducted an experiment on r/changemyview using AI-generated comments
- The experiment operated under IRB approval 24.04.01
- The r/changemyview moderators filed an ethics complaint
- The principal investigator received a formal warning
- The experiment involved personalized persuasion tactics based on user posting histories
Unclear:
- The duration of the experiment and number of AI-generated comments posted
- The specific language models and technical systems employed
- Whether any users successfully had their views changed by AI-generated arguments
- The full terms of the formal warning and any additional sanctions
- Whether the research findings will be published or suppressed
- The identities of the researchers involved
What to Watch Next
The University of Zurich may issue additional statements clarifying the scope of the experiment and the institutional response. Academic ethics bodies may use this case to develop updated guidelines for AI research involving human subjects on public platforms.
Reddit's response to the incident could include policy changes regarding researcher access or enhanced detection mechanisms for AI-generated content. The platform has not issued a public statement as of the time of this report.
The r/changemyview community may implement new verification procedures or rules designed to prevent similar experiments. The moderators indicated they are reviewing their policies in light of the incident.
Academic institutions conducting AI research may face increased scrutiny of their IRB processes for experiments involving AI systems that interact with human subjects. The case could prompt revisions to research ethics guidelines at universities worldwide.

