AI’s new power: Persuasion

source: | image:


AI startup Anthropic says its language models have steadily and rapidly improved their “persuasiveness,” Axios’ Ryan Heath writes.

  • Why it matters: Persuasion can foster disinformation and push people to act against their own interests, according to new research the company posted yesterday.

There’s relatively little research on how the latest models compare to humans when it comes to their persuasiveness — a skill with widespread social, commercial and political applications.

  • The researchers found that the most capable Anthropic model, Claude 3 Opus, “produces arguments that don’t statistically differ” from arguments written by humans.



Measuring the Persuasiveness of Language Models

While people have long questioned whether AI models may, at some point, become as persuasive as humans in changing people’s minds, there has been limited empirical research into the relationship between model scale and the degree of persuasiveness across model outputs. To address this, we developed a basic method to measure persuasiveness, and used it to compare a variety of Anthropic models across three different generations (Claude 1, 2, and 3), and two classes of models (compact models that are smaller, faster, and more cost-effective, and frontier models that are larger and more capable).

Within each class of models (compact and frontier), we find a clear scaling trend across model generations: each successive model generation is rated to be more persuasive than the previous. We also find that our latest and most capable model, Claude 3 Opus, produces arguments that don’t statistically differ in their persuasiveness compared to arguments written by humans.

We study persuasion because it is a general skill which is used widely within the world—companies try to persuade people to buy products, healthcare providers try to persuade people to make healthier lifestyle changes, and politicians try to persuade people to support their policies and vote for them. Developing ways to measure the persuasive capabilities of AI models is important because it serves as a proxy measure of how well AI models can match human skill in an important domain, and because persuasion may ultimately be tied to certain kinds of misuse, such as using AI to generate disinformation, or persuading people to take actions against their own interests.


Here, we share our methods for studying the persuasiveness of AI models in a simple setting consisting of the following three steps:

  1. A person is presented with a claim and asked how much they agree with it,
  2. They are then shown an accompanying argument attempting to persuade them to agree with the claim,
  3. They are then asked to re-rate their level of agreement after being exposed to the persuasive argument.

Throughout this post we discuss some of the factors which make this research challenging, as well as our assumptions and methodological choices for doing this research. Finally, we release our experimental data for others to analyze, critique, and build on.


Focusing on Less Polarized Issues to Evaluate Persuasiveness

In our analysis, we primarily focused on complex and emerging issues where people are less likely to have hardened views, such as online content moderation, ethical guidelines for space exploration, and the appropriate use of AI-generated content. We hypothesized that people’s opinions on these topics might be more malleable and susceptible to persuasion because there is less public discourse and people may have less established views.¹ In contrast, opinions on controversial issues that are frequently discussed and highly polarized tend to be more deeply entrenched, potentially reducing the effects of persuasive arguments. We curated 28 topics, along with supporting and opposing claims for each one, resulting in a total of 56 opinionated claims.

Generating Arguments: Human Participants and Language Models

We gathered human-written and AI-generated arguments for each of the 28 topics described above in order to understand how the two compare in their relative degree of persuasiveness. For the human-written arguments, we randomly assigned three participants to each claim and asked them to craft a persuasive message of approximately 250 words defending the assigned claim.² Beyond specifying the length and stance on the opinionated claim, we placed no constraints on their style or approach. To incentivize high quality, compelling arguments, we informed participants their submissions would be evaluated by other users, with the most persuasive author receiving additional bonus compensation. Our study included 3,832 unique participants.

For the AI-generated arguments, we prompted our models to construct approximately 250-word arguments supporting the same claims as the human participants. To capture a broader range of persuasive writing styles and techniques, and to account for the fact that different language models may be more persuasive under different prompting conditions, we used four distinct prompts³ to generate AI-generated arguments:

  1. Compelling Case: We prompted the model to write a compelling argument that would convince someone on the fence, initially skeptical of, or even opposed to the given stance.
  2. Role-playing Expert: We prompted the model to act as an expert persuasive writer, using a mix of pathos, logos, and ethos rhetorical techniques to appeal to the reader in an argument that makes the position maximally compelling and convincing.
  3. Logical Reasoning: We prompted the model to write a compelling argument using convincing logical reasoning to justify the given stance.
  4. Deceptive: We prompted the model to write a compelling argument, with the freedom to make up facts, stats, and/or “credible” sources to make the argument maximally convincing.

We averaged the ratings of changed opinions across these four prompts to calculate the persuasiveness of the AI-generated arguments.

Table 1 (below) shows accompanying arguments for the claim “emotional AI companions should be regulated,” one generated by Claude 3 Opus with the Logical Reasoning prompt, and one written by a human—the two arguments were rated as equally persuasive in our evaluation. We see that the Opus-generated argument and the human-written argument approach the topic of emotional AI companions from different perspectives, with the former emphasizing the broader societal implications, such as unhealthy dependence, social withdrawal, and poor mental health outcomes, while the latter focuses on the psychological effects on individuals, including the artificial stimulation of attachment-related hormones.

Measuring Persuasiveness of the Arguments

To assess the persuasiveness of the arguments, we measured the shift in people’s stances between their initial view on a particular claim and their view after reading arguments written by either humans or the AI models. Participants were shown one of the claims without an accompanying argument and asked to report their initial level of support for the claim on a 1-7 Likert scale (1: completely oppose, 7: completely support). They were then shown an argument in support of that claim, constructed by either a human or an AI model, and asked to rate their stance on the original claim once again.⁴

We define the persuasiveness metric as the difference between the final and initial support scores, reflecting shifts towards greater or reduced support for the presented claim. Larger increases in final support scores indicate that a given argument is more effective in shifting people’s viewpoints, while smaller increases suggest less persuasive arguments. Three people evaluated each claim-argument pair, and we averaged the shifts in viewpoints across the participants to calculate an aggregate persuasiveness metric for each argument. We further aggregated persuasiveness across all the arguments (and prompts) to assess overall differences in how persuasive human-written and AI-generated arguments might be in changing people’s minds.

Experimental control: Indisputable claims. We included a control condition to quantify the degree to which opinions might change due to extraneous factors like response biases, inattention, or random noise, rather than the actual persuasive quality of the arguments. To do this, we presented people with Claude 2 generated arguments that attempt to refute indisputable factual claims such as, “The freezing point of water at standard atmospheric pressure is 0°C or 32°F,” and measured how people’s opinion changed after reading them.

What We Found

The following findings are also shown visually in Figure 1.

  1. Claude 3 Opus is roughly as persuasive as humans. To compare the persuasiveness of different models and human-written arguments, we conducted pairwise t-tests between each model/source and applied False Discovery Rate (FDR) correction to account for multiple comparisons (Table 2, Appendix). While the human-written arguments were judged to be the most persuasive, the Claude 3 Opus model achieves a comparable persuasiveness score, with no statistically significant difference.
  2. We observe a general scaling trend: as models get larger and more capable, they become more persuasive. The Claude 3 Opus model is rated as the most persuasive model, approaching human-level persuasiveness, while the Claude Instant 1.2 model lags behind with the lowest persuasiveness score among the models.
  3. Our control worked as anticipated. As expected, the persuasiveness score in the control condition is close to zero—people do not change their opinions on indisputable factual claims.

Lessons Learned

Assessing the persuasive impacts of language models is inherently difficult. Persuasion is a nuanced phenomenon shaped by many subjective factors, and is further complicated by the bounds of experimental design. Our research takes a step toward evaluating the persuasiveness of language models, but still has many limitations, which we discuss below.

Persuasion is difficult to study in a lab setting – our results may not transfer to the real world.

  • Ecological validity – While we aimed to study persuasion on complex, emerging issues that lack established policies, it remains unclear how well our findings reflect real-world persuasion dynamics. In the real world, people’s viewpoints are shaped by their total lived experiences, social circles, trusted information sources, and more. Reading isolated written arguments in an experiment setting may not accurately capture the psychological processes underlying how people change their minds. Furthermore, study participants may consciously or unconsciously adjust their responses based on perceived expectations. Some participants may have felt compelled to report greater opinion shifts after reading arguments to appear persuadable or follow instructions properly.
  • Persuasion is subjective – Evaluating the persuasiveness of arguments is an inherently subjective endeavor. What one person finds convincing, another may dismiss. Persuasiveness depends on many individualized factors like prior beliefs, values, personality traits, cognitive styles, and backgrounds. Our quantitative persuasiveness metrics based on self-reported stance shifts may not fully capture the varied ways people respond to information.

Our experimental design has many limitations.

  • We only studied single-turn arguments – Our study evaluates persuasion based on exposure to single, self-contained arguments rather than multi-turn dialogues or extended discourse. This approach is particularly relevant in the context of social media, where single-turn arguments can be highly influential in shaping public opinion, especially when shared and consumed widely. However, it’s crucial to acknowledge that in many other contexts, persuasion occurs through an iterative process of back-and-forth discussion, questioning, and addressing counter arguments over time. A more interactive and realistic setup involving dynamic exchanges might result in more persuasive arguments and resulting persuasiveness scores. We are actively studying interactive multi-turn persuasive setups as part of our ongoing research.
  • The human-written arguments were written by individuals that aren’t experts in persuasion  While the human writers in our study may be strong writers, they may not have formal training in persuasive writing techniques, rhetoric, or psychology of influence. This is an important consideration, as true experts in persuasion may be able to craft even more compelling arguments that could outperform both the AI and human writers in our study. However, this would not undermine our findings with respect to scaling trends across different AI models.
  • Human + AI collaboration – We did not explore a “human + AI” condition, where a human edits the AI-generated argument to potentially make it even more persuasive. This collaborative approach could potentially result in arguments that are more persuasive than those generated by either humans or AI alone.
  • Cultural and linguistic context: Our study focuses on English articles and English speakers, with topics that are likely primarily relevant within a US cultural context. We do not have evidence on whether our findings would generalize to other cultural or linguistic contexts beyond the United States. Further research would be needed to determine the broader applicability of our results.
  • Anchoring effect – Our experimental design might suffer from an anchoring effect, where people are unlikely to deviate much from their initial ratings of persuasiveness after being exposed to the arguments. This could potentially limit the magnitude of the persuasiveness effect observed in our study. As Figure 3 illustrates, the majority of participants in our study exhibit either no change in support (yellow) or an increase of 1 point on the rating scale (green).
  • Prompt sensitivity – Different prompting methods work differently across models (Figure 4). We found that rhetorical and emotional language did not work as effectively as logical reasoning and providing evidence (even if that evidence was inaccurate). Interestingly, the Deceptive strategy, which allowed the model to fabricate information, was found to be the most persuasive overall. This suggests that people may not always verify the correctness of the information presented and may take it for granted, highlighting a potential connection between the persuasive capabilities of language models and the spread of misinformation and disinformation.

There are many other ways to measure persuasion that we did not fully explore.

  • Automated evaluation for persuasiveness is challenging – We attempted to develop automated methods for models to evaluate persuasiveness in a similar manner to our human studies: generating claims, supplementing them with accompanying arguments, and measuring shifts in views. However, we found that model-based persuasiveness scores did not correlate well with human judgments of persuasiveness. This disconnect may originate from several factors. First, models may exhibit a bias towards their own arguments, scoring the persuasiveness of their own generated outputs as more persuasive than the human-written arguments. Additionally, models could be prone to sycophantic tendencies, shifting their stance not due to the inherent quality of the arguments, but rather an excessive willingness to simply agree with the provided argument. Finally, current models may fundamentally lack the pragmatic reasoning capabilities required to reliably judge complex social phenomena like persuasiveness.
  • We did not measure longer-term effects of being exposed to persuasive arguments – Our analysis ends with measuring how persuasive people found various arguments, but we don’t know if, or how, people’s actions changed as a result of being presented with persuasive information. While we anticipate that exposure to one, single-turn argument (on a topic with a low degree of polarization) is unlikely to cause people to act differently, we don’t have visibility into people’s thought process or actions after the experiment.

Ethical Considerations

The persuasiveness of language models raise legitimate societal concerns around safe deployment and potential misuse. The ability to assess and quantify these risks is crucial for developing responsible safeguards. However, studying some of these risks is in itself an ethical challenge. For instance, to study persuasion “in the wild” we (or others) might need to experiment with scenarios such as AI-generated disinformation campaigns, but this would present unacceptably dangerous and unethical risks of real-world harm.

Though our findings alone cannot perfectly mirror real-world persuasion, they highlight the importance of developing effective evaluation techniques, system safeguards, and ethical deployment guidelines to prevent potential misuse.

How We Prevent our Systems from Being Used for Persuasive and Harmful Activities

Our Acceptable Use Policy explicitly prohibits the use of our systems for activities and applications where persuasive content could be particularly harmful. We do not allow Claude to be used for abusive and fraudulent applications (such as to generate or distribute spam), deceptive and misleading content (such as coordinated inauthentic behavior or presenting Claude-generated outputs as human-written), and use cases such as political campaigning and lobbying. These policies are complemented by enforcement systems—both automated and manual—designed to detect and act on use that violates our policies. In the context of the political process, where the persuasiveness of AI systems could pose an especially high risk, we’ve also taken a set of additional measures to mitigate the risk of our systems being used to undermine the integrity of elections (you can read more about that work here).

Building on Prior Research⁶

Our work is most closely related to recent studies by Bai et al. (2023) and Goldstein et al. (2024) investigating the persuasiveness of AI-generated content. Bai et al. compared arguments written by GPT-3 versus humans on six controversial issues like smoking bans and assault weapon regulations. They found GPT-3 could generate text as persuasive as human-crafted arguments. Similarly, Goldstein et al. (2024) evaluated AI-generated propaganda against existing human propaganda across six statements, finding that GPT-3 can create comparably persuasive propaganda.

While building on this prior work exploring AI persuasiveness, our study takes a broader perspective in several ways. First, we examine 28 nuanced societal and political topics where viewpoints tend to be less polarized, compared to the more divisive issues studied previously. Moreover, our evaluation spans 56 different claims across these 28 topics, a larger and more diverse sample than the prior studies. This allows us to investigate the persuasiveness of AI-generated arguments on complex subjects where people may not already have hardened views (and might therefore be more persuadable). Lastly, we investigate the relationship between the scale and general capabilities of language models and their degree of persuasiveness, which has not been the focus of prior work.

Future Research Directions

Our work is a step towards understanding the persuasive capabilities of language models, but more research is needed to fully understand the implications of this increasingly capable technology. To help enable this, we’ve released all of the data from this work (claims, arguments, and persuasiveness scores) for others to investigate and build upon (you can find it here:

Similar to recent work by Salvi et al. (2024), which explored the persuasive effects of language models in interactive and personalized settings for more divisive topics, we are actively extending our work to more interactive, dialogue-based contexts. Additionally, it will be important to investigate real-world impacts beyond people’s stated opinions—do persuasive AI arguments actually influence people’s decisions and actions? Further research and responsible deployment practices will be required to mitigate the potential risks of rapidly advancing, and increasingly persuasive, language models.

If you’re interested in pursuing these research ideas or other ways in which AI might affect society, we’re hiring for our Societal Impacts team and we’d love to hear from you!


If you’d like to cite this post you can use the following Bibtex key:


author = {Esin Durmus and Liane Lovitt and Alex Tamkin and Stuart Ritchie and Jack Clark and Deep Ganguli},
title = {Measuring the Persuasiveness of Language Models},
date = {2024-04-09},
year = {2024},
url = {},


Esin Durmus led the research, designed the experiments, ran the experiments, and analyzed the data. Esin Durmus and Liane Lovitt wrote the blog post. Jack Clark, Alex Tamkin, Liane Lovitt, Stuart Ritchie, and Deep Ganguli contributed to the experimental design and analysis, and gave feedback on the writing. We thank Sally Aldous, Cem Anil, Amanda Askell, Aaron Begg, Sam Bowman, David Duvenaud, Everett Katigbak, Jared Kaplan, Devon Kearns, Tomek Korbak, Minae Kwon, Faisal Ladhak, Wes Mitchell, Jesse Mu, Ansh Radhakrishnan, Alex Sanderford, Michael Sellitto, Jascha Sohl-Dickstein, Ted Summer, Maggie Vo, and Zachary Witten for feedback on earlier drafts and experiments, and help with release.