ParCIS Lab, BUPT (ParCIS)

ParCIS Lab, BUPT

ParCIS

Organization data from Github https://github.com/ParCIS

Parallel Computing and Intelligent Systems Laboratory (ParCIS Lab), Beijing University of Posts and Telecommunications

Location:China

Home Page:https://shigangli.github.io/

GitHub:@ParCIS

ParCIS Lab, BUPT's repositories

Magicube

Magicube is a high-performance library for quantized sparse matrix operations (SpMM and SDDMM) of deep learning on Tensor Cores.

Language:C++License:GPL-3.0Stargazers:89Issues:3Issues:2

Chimera

Chimera: bidirectional pipeline parallelism for efficiently training large-scale models.

Language:PythonLicense:GPL-3.0Stargazers:67Issues:1Issues:4

FlashSparse

FlashSparse significantly reduces the computation redundancy for unstructured sparsity (for SpMM and SDDMM) on Tensor Cores through a Swap-and-Transpose mapping strategy. FlashSparse is accepted by PPoPP 2025.

Language:CudaStargazers:29Issues:1Issues:0

Ok-Topk

Ok-Topk is a scheme for distributed training with sparse gradients. Ok-Topk integrates a novel sparse allreduce algorithm (less than 6k communication volume which is asymptotically optimal) with the decentralized parallel Stochastic Gradient Descent (SGD) optimizer, and its convergence is proved theoretically and empirically.

Language:PythonLicense:GPL-3.0Stargazers:27Issues:1Issues:4

DNN-cpp-proxies

C++/MPI proxies for distributed training of deep neural networks.

Language:C++License:GPL-3.0Stargazers:1Issues:2Issues:0