How Johnny Can Persuade LLMs to Jailbreak Them:<br>Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs
- Paper
- Jan 12, 2024
- #LLM #ArtificialIntelligence #Persuasion
We study how to persuade LLMs to jailbreak them and advocate for more fundamental mitigation for highly interactive LLMs