Login / Signup
MPress: Democratizing Billion-Scale Model Training on Multi-GPU Servers via Memory-Saving Inter-Operator Parallelism.
Quan Zhou
Haiquan Wang
Xiaoyan Yu
Cheng Li
Youhui Bai
Feng Yan
Yinlong Xu
Published in:
HPCA (2023)
Keyphrases
</>
high level
computational model
probabilistic model
management system
supervised learning
prior knowledge
conceptual model
multiscale
mathematical model
formal model
data sets
similarity measure
probability distribution
training samples
experimental data