Login / Signup

Prefixing Attention Sinks can Mitigate Activation Outliers for Large Language Model Quantization.

Seungwoo SonWonpyo ParkWoohyun HanKyuyeun KimJaeho Lee
Published in: CoRR (2024)
Keyphrases