Explanation Guided Knowledge Distillation for Pre-trained Language Model Compression.
Zhao YangYuanzhe ZhangDianbo SuiYiming JuJun ZhaoKang LiuPublished in: ACM Trans. Asian Low Resour. Lang. Inf. Process. (2024)
Keyphrases
- language model
- language modeling
- pre trained
- n gram
- document retrieval
- speech recognition
- probabilistic model
- retrieval model
- language modelling
- test collection
- information retrieval
- query expansion
- ad hoc information retrieval
- neural network
- hidden markov models
- image retrieval
- training data
- face recognition
- feature selection