OWQ: Outlier-Aware Weight Quantization for Efficient Fine-Tuning and Inference of Large Language Models.
Changhun LeeJungyu JinTaesu KimHyungjun KimEunhyeok ParkPublished in: AAAI (2024)
Keyphrases
- language model
- fine tuning
- language modeling
- n gram
- language modelling
- document retrieval
- test collection
- probabilistic model
- query expansion
- speech recognition
- statistical language models
- retrieval model
- context sensitive
- language model for information retrieval
- smoothing methods
- information retrieval
- relevance model
- weighting scheme
- viable alternative
- cross lingual
- collaborative filtering
- hidden markov models