Persuasion
Title: The large-scale collaborative AI Persuasion Project has a website now! Our project is a large-scale collaborative effort to study persuasion capabilities of frontier Large Language Models (LLMs). Our team consists of social scientists with a background in LLM research, AI researchers, as well as a host of additional and diverse profiles (such as linguistics, communication, engineering, etc). We have started designing experiments in early April 2024.
The primary goal of this project is to provide a general, detailed, and extensive assessment of LLM persuasion capabilities in a high-external-validity context. This work includes evaluations of model capabilities, showing potential effects across model sizes and prompting efforts, as well as between languages.
Goal and scope
Specifically, at present, our project is investigating LLM persuasion performance in an abstract and general quiz setting, where we compare LLM performance compared to human persuaders, replicating these results across models and languages like Chinese.
This project aims to deliver a strong understanding of current persuasion capabilities as well as a framework that can be applied readily and quickly once future iterations of frontier models are released. This continual updating of our research thus provides a repeatable assessment of persuasion capabilities that allows decision makers at different points in time to accurately estimate LLM risks that may arise from persuasion.
Head over there for information about what we're up to, our team, and (eventually) our outputs as they come out!