Sign in

How Johnny Can Persuade LLMs to Jailbreak Them: Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs.

Yi ZengHongpeng LinJingwen ZhangDiyi YangRuoxi JiaWeiyan Shi
Published in: CoRR (2024)
Keyphrases