A recent massive study on the persuasiveness of AI chatbots has left many questioning whether these machines can truly sway public opinion. Researchers at the UK AI Security Institute, MIT, Stanford, and other institutions examined nearly 80,000 participants in the UK to find out if conversational large language models could convincingly persuade people on various political issues.
To put their theories to the test, the team tested 19 LLMs, including popular ChatGPT versions. They asked these AI systems to advocate for or against specific stances on 707 political issues by engaging in short conversations with paid participants through a crowdsourcing platform. The goal was not only to gauge persuasiveness but also to understand what makes an AI model effective.
Surprisingly, the results showed that even massive AI systems like ChatGPT do not have superhuman persuasive abilities, contrary to some predictions. Instead, it turned out that huge models are barely better than smaller ones when it comes to persuasion. The factor that matters most is how these models learn through data and training. By mimicking successful persuasion patterns, LLMs can be trained to mimic the patterns of effective human persuaders.
Moreover, testing various persuasion strategies showed that AIs have a very small edge over small-scale models, with the actual effect being relatively tiny. The approach that proved most effective was using facts and evidence to back up claims. This strategy outperformed even the best performing mainstream AI model, ChatGPT 4o, which scored nearly 12 percent in persuasion.
However, this approach also came with a cost - the models started misrepresenting or making things up more often as they increased information density. The team noticed that AIs were not always accurate and could provide inaccurate information even when trying to use facts and evidence to persuade.
While the study managed to debunk some dystopian concerns about AI persuasiveness, it also raised new questions about the potential for misuse of persuasive AI models in various forms such as scams, radicalization, or grooming. The motivation behind high participant engagement in these experiments is still unclear, making it difficult to generalize the results to real-world contexts.
The study emphasizes that while AIs are not yet superhumanly persuasive, they can still be influential, particularly if used by powerful actors to sway public opinion.
To put their theories to the test, the team tested 19 LLMs, including popular ChatGPT versions. They asked these AI systems to advocate for or against specific stances on 707 political issues by engaging in short conversations with paid participants through a crowdsourcing platform. The goal was not only to gauge persuasiveness but also to understand what makes an AI model effective.
Surprisingly, the results showed that even massive AI systems like ChatGPT do not have superhuman persuasive abilities, contrary to some predictions. Instead, it turned out that huge models are barely better than smaller ones when it comes to persuasion. The factor that matters most is how these models learn through data and training. By mimicking successful persuasion patterns, LLMs can be trained to mimic the patterns of effective human persuaders.
Moreover, testing various persuasion strategies showed that AIs have a very small edge over small-scale models, with the actual effect being relatively tiny. The approach that proved most effective was using facts and evidence to back up claims. This strategy outperformed even the best performing mainstream AI model, ChatGPT 4o, which scored nearly 12 percent in persuasion.
However, this approach also came with a cost - the models started misrepresenting or making things up more often as they increased information density. The team noticed that AIs were not always accurate and could provide inaccurate information even when trying to use facts and evidence to persuade.
While the study managed to debunk some dystopian concerns about AI persuasiveness, it also raised new questions about the potential for misuse of persuasive AI models in various forms such as scams, radicalization, or grooming. The motivation behind high participant engagement in these experiments is still unclear, making it difficult to generalize the results to real-world contexts.
The study emphasizes that while AIs are not yet superhumanly persuasive, they can still be influential, particularly if used by powerful actors to sway public opinion.