skip to content
LLM Persuasion Safety Hub

Welcome

LLM Persuasion Safety Hub is a collection of datasets and codebases used to study and evaluate persuasion in large language models (LLMs). The goal is to bring together the most relevant resources as they emerge, making safety-relevant LLM persuasion research easier to discover, compare, and reproduce.

This initiative grew out of work on a survey of automatic persuasion evaluation methods: Measuring Machine Persuasion: A Survey of Automated Evaluation Methods for Large Language Models (work in progress), poster to be presented at the first AIMII workshop at IASEAI'26.

What is included

How to contribute

If you know of a relevant resource that is missing, there are two ways to add it: