Login / Signup

Rethinking Kullback-Leibler Divergence in Knowledge Distillation for Large Language Models.

Taiqiang WuChaofan TaoJiahao WangZhe ZhaoNgai Wong
Published in: CoRR (2024)
Keyphrases