JY-Ren's repositories
transpeeder
train llama on a single A100 80G node using 🤗 transformers and 🚀 Deepspeed Pipeline Parallelism
neox-LLAMA2
update neox framework to train llama1&llama2
Language:PythonApache-2.0000
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Apache-2.0000