​
Login / Signup
Keshi Ge
ORCID
Publication Activity (10 Years)
Years Active: 2018-2024
Publications (10 Years): 16
Top Topics
Graph Classification
Distributed Data
Deep Learning
Load Balancing
Top Venues
CLUSTER
CoRR
IEEE Trans. Parallel Distributed Syst.
IEEE J. Sel. Areas Commun.
</>
Publications
</>
Shengwei Li
,
Kai Lu
,
Zhiquan Lai
,
Weijie Liu
,
Keshi Ge
,
Dong Sheng Li
A Multidimensional Communication Scheduling Method for Hybrid Parallel DNN Training.
IEEE Trans. Parallel Distributed Syst.
35 (8) (2024)
Zhiquan Lai
,
Shengwei Li
,
Xudong Tang
,
Keshi Ge
,
Weijie Liu
,
Yabo Duan
,
Linbo Qiao
,
Dongsheng Li
Merak: An Efficient Distributed DNN Training Framework With Automated 3D Parallelism for Giant Foundation Models.
IEEE Trans. Parallel Distributed Syst.
34 (5) (2023)
Wei Wang
,
Zhiquan Lai
,
Shengwei Li
,
Weijie Liu
,
Keshi Ge
,
Yujie Liu
,
Ao Shen
,
Dongsheng Li
Prophet: Fine-grained Load Balancing for Parallel Training of Large-scale MoE Models.
CLUSTER
(2023)
Keshi Ge
,
Kai Lu
,
Yongquan Fu
,
Xiaoge Deng
,
Zhiquan Lai
,
Dongsheng Li
Compressed Collective Sparse-Sketch for Distributed Data-Parallel Training of Deep Learning Models.
IEEE J. Sel. Areas Commun.
41 (4) (2023)
Hongyu Chen
,
Zhejiang Ran
,
Keshi Ge
,
Zhiquan Lai
,
Jingfei Jiang
,
Dongsheng Li
Auto-Divide GNN: Accelerating GNN Training with Subgraph Division.
Euro-Par
(2023)
Shengwei Li
,
Zhiquan Lai
,
Yanqi Hao
,
Weijie Liu
,
Keshi Ge
,
Xiaoge Deng
,
Dongsheng Li
,
Kai Lu
Automated Tensor Model Parallelism with Overlapped Communication for Efficient Foundation Model Training.
CoRR
(2023)
Keshi Ge
,
Zhejiang Ran
,
Zhiquan Lai
,
Lizhi Zhang
,
Dongsheng Li
BRGraph: An efficient graph neural network training system by reusing batch data on GPU.
Concurr. Comput. Pract. Exp.
34 (15) (2022)
Zhiquan Lai
,
Shengwei Li
,
Xudong Tang
,
Keshi Ge
,
Weijie Liu
,
Yabo Duan
,
Linbo Qiao
,
Dongsheng Li
Merak: An Efficient Distributed DNN Training Framework with Automated 3D Parallelism for Giant Foundation Models.
CoRR
(2022)
Weijie Liu
,
Zhiquan Lai
,
Shengwei Li
,
Yabo Duan
,
Keshi Ge
,
Dongsheng Li
AutoPipe: A Fast Pipeline Parallelism Approach with Balanced Partitioning and Micro-batch Slicing.
CLUSTER
(2022)
Yabo Duan
,
Zhiquan Lai
,
Shengwei Li
,
Weijie Liu
,
Keshi Ge
,
Peng Liang
,
Dongsheng Li
HPH: Hybrid Parallelism on Heterogeneous Clusters for Accelerating Large-scale DNNs Training.
CLUSTER
(2022)
Keshi Ge
,
Yongquan Fu
,
Yiming Zhang
,
Zhiquan Lai
,
Xiaoge Deng
,
Dongsheng Li
S2 Reducer: High-Performance Sparse Communication to Accelerate Distributed Deep Learning.
ICASSP
(2022)
Keshi Ge
,
Yongquan Fu
,
Zhiquan Lai
,
Xiaoge Deng
,
Dongsheng Li
S2 Reducer: High-Performance Sparse Communication to Accelerate Distributed Deep Learning.
CoRR
(2021)
Keshi Ge
,
Yiming Zhang
,
Yongquan Fu
,
Zhiquan Lai
,
Xiaoge Deng
,
Dongsheng Li
CASQ: Accelerate Distributed Deep Learning with Sketch-Based Gradient Quantization.
CLUSTER
(2021)
Yixin Chen
,
Xinye Lin
,
Keshi Ge
,
Wenbo He
,
Dongsheng Li
Tag Pollution Detection in Web Videos via Cross-Modal Relevance Estimation.
IWQoS
(2020)
Dongsheng Li
,
Zhiquan Lai
,
Keshi Ge
,
Yiming Zhang
,
Zhaoning Zhang
,
Qinglin Wang
,
Huaimin Wang
HPDL: Towards a General Framework for High-performance Distributed Deep Learning.
ICDCS
(2019)
Lei Guan
,
Linbo Qiao
,
Dongsheng Li
,
Tao Sun
,
Keshi Ge
,
Xicheng Lu
An Efficient ADMM-Based Algorithm to Nonconvex Penalized Support Vector Machines.
ICDM Workshops
(2018)