Login / Signup
MiniLMv2: Multi-Head Self-Attention Relation Distillation for Compressing Pretrained Transformers.
Wenhui Wang
Hangbo Bao
Shaohan Huang
Li Dong
Furu Wei
Published in:
ACL/IJCNLP (Findings) (2021)
Keyphrases
</>
visual attention
data compression
data sets
databases
data structure
viewpoint
database
real world
image sequences
face detection
focus of attention