Login / Signup

MLKD-BERT: Multi-level Knowledge Distillation for Pre-trained Language Models.

Ying ZhangZiheng YangShufan Ji
Published in: CoRR (2024)
Keyphrases