Sign in

Layer-wise Pruning of Transformer Attention Heads for Efficient Language Modeling.

Kyuhong ShimIksoo ChoiWonyong SungJungwook Choi
Published in: ISOCC (2021)
Keyphrases