Login / Signup

Revisiting Knowledge Distillation for Autoregressive Language Models.

Qihuang ZhongLiang DingLi ShenJuhua LiuBo DuDacheng Tao
Published in: CoRR (2024)
Keyphrases