Login / Signup
MiniLMv2: Multi-Head Self-Attention Relation Distillation for Compressing Pretrained Transformers.
Wenhui Wang
Hangbo Bao
Shaohan Huang
Li Dong
Furu Wei
Published in:
CoRR (2020)
Keyphrases
</>
real time
data sets
artificial intelligence
case study
real world
information retrieval
search engine
computer vision
rough sets
data compression
head pose estimation
partial discharge