Zhenghui Jin's repositories
incubator-mxnet
Lightweight, Portable, Flexible Distributed/Mobile Deep Learning with Dynamic, Mutation-aware Dataflow Dep Scheduler; for Python, R, Julia, Scala, Go, Javascript and more
alpa
Auto parallelization for large-scale neural networks
amazon-eks-ami
Packer configuration for building a custom EKS AMI
apex
A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch
array-api-tests
Test suite for the PyData Array APIs standard
arrow
Apache Arrow is a multi-language toolbox for accelerated data interchange and in-memory processing
aws-efa-eks
Deploying P4d in EKS utilizing GPUDirectRDMA over EFA
aws-efa-nccl-baseami-pipeline
EFA/NCCL base AMI build Packer and CodeBuild/Pipeline files. Also base Docker build files to enable EFA/NCCL in containers
d2l-en
Interactive deep learning book with multi-framework code, math, and discussions. Adopted at 175 universities.
dcgm-exporter
NVIDIA GPU metrics exporter for Prometheus leveraging DCGM
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.
elastic
PyTorch elastic training
gluon-nlp
NLP made easy
horovod
Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.
jax
Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
kaldi
kaldi-asr/kaldi is the official location of the Kaldi project.
mxnet_recipes
MXNet conda recipes
nccl
Optimized primitives for collective multi-GPU communication
NeMo
NeMo: a toolkit for conversational AI
python-etcd
A python client for etcd
pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
pytorch-lightning
The lightweight PyTorch wrapper for high-performance AI research. Scale your models, not the boilerplate.
ray_lightning
Pytorch Lightning Distributed Accelerators using Ray
staged-recipes
A place to submit conda recipes before they become fully fledged conda-forge feedstocks
torchx
TorchX is a library containing standard DSLs for authoring and running PyTorch related components for an E2E production ML pipeline
training-operator
Training operators on Kubernetes.
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
xformers
Hackable and optimized Transformers building blocks, supporting a composable construction.
xla
Enabling PyTorch on XLA Devices (e.g. Google TPU)