Login / Signup

Knowledge Distillation as Efficient Pre-training: Faster Convergence, Higher Data-efficiency, and Better Transferability.

Ruifei HeShuyang SunJihan YangSong BaiXiaojuan Qi
Published in: CoRR (2022)
Keyphrases