ACM CHI 2025 · Research Publication
April 2025
In a randomized controlled trial study, “Deceptive Explanations by Large Language Models Lead People to Change Their Beliefs About Misinformation More Often than Honest Explanations,” examines how AI systems affect human beliefs when they provide inaccurate reasoning via explanations alongside their outputs. We find that not only can deceptive reasoning make people believe false information more than they already do without AI, but deceptive reasoning by an AI chatbot can even be more persuasive than when AI provides accurate reasoning. As AI systems are increasingly capable of providing strong reasoning to persuade people and can even do so with hidden motives unknown to the AI platform developers, this work underscores their potential to inadvertently amplify misinformation, shape public opinion in unintended ways, and exacerbate the societal risks associated with automated persuasion.
Our Approach
In a preregistered online experiment, nearly 600 participants evaluated the truth of a series of true and fake news titles. Participants were then provided with feedback from either a deceptive or accurate AI in one of two formats:
- Without Explanations: The AI system simply states if a piece of information is true or false.
- With Explanations: A classification accompanied by an AI generated explanation of why something is true or false.
This setup allowed us to isolate the influence of the AI-provided explanations on belief revision.