Don’t Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models
Published in USENIX Security Symposium, 2024
This paper is about investigating LLM jailbreak threats, from the perspective of empirical evaluation and automatic generation.
Download here