Login / Signup
Rethinking Attention: Exploring Shallow Feed-Forward Neural Networks as an Alternative to Attention Layers in Transformers.
Vukasin Bozic
Danilo Dordevic
Daniele Coppola
Joseph Thommes
Published in:
CoRR (2023)
Keyphrases
</>
feed forward neural networks
neural network
multi layer
machine learning
training data
artificial neural networks
non stationary
genetic algorithm
data fusion
feed forward