Login / Signup

Linear attention is (maybe) all you need (to understand transformer optimization).

Kwangjun AhnXiang ChengMinhak SongChulhee YunAli JadbabaieSuvrit Sra
Published in: CoRR (2023)
Keyphrases