Login / Signup

One-Shot Safety Alignment for Large Language Models via Optimal Dualization.

Xinmeng HuangShuo LiEdgar DobribanOsbert BastaniHamed HassaniDongsheng Ding
Published in: CoRR (2024)
Keyphrases