Login / Signup
How Much Does Attention Actually Attend? Questioning the Importance of Attention in Pretrained Transformers.
Michael Hassid
Hao Peng
Daniel Rotem
Jungo Kasai
Ivan Montero
Noah A. Smith
Roy Schwartz
Published in:
EMNLP (Findings) (2022)
Keyphrases
</>
computer vision
pattern recognition
data analysis
focus of attention
information retrieval
genetic algorithm
feature selection
website
training data
cooperative
support vector
preprocessing
information technology
hidden markov models