Yimin Jiang (ymjiang)

ymjiang

Geek Repo

Company:AML@ByteDance

Github PK Tool:Github PK Tool


Organizations
bytedance
dmlc

Yimin Jiang's starred repositories

S-LoRA

S-LoRA: Serving Thousands of Concurrent LoRA Adapters

Language:PythonLicense:Apache-2.0Stargazers:1709Issues:0Issues:0

MS-AMP

Microsoft Automatic Mixed Precision Library

Language:PythonLicense:MITStargazers:510Issues:0Issues:0

Medusa

Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Heads

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:2216Issues:0Issues:0

flash-attention

Fast and memory-efficient exact attention

Language:PythonLicense:BSD-3-ClauseStargazers:13504Issues:0Issues:0

trlx

A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)

Language:PythonLicense:MITStargazers:4451Issues:0Issues:0

Prompt-Engineering-Guide

šŸ™ Guides, papers, lecture, notebooks and resources for prompt engineering

Language:MDXLicense:MITStargazers:47893Issues:0Issues:0

nanoGPT

The simplest, fastest repository for training/finetuning medium-sized GPTs.

Language:PythonLicense:MITStargazers:36327Issues:0Issues:0
Language:PythonStargazers:212Issues:0Issues:0

PiPPy

Pipeline Parallelism for PyTorch

Language:PythonLicense:BSD-3-ClauseStargazers:712Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:205Issues:0Issues:0

pcm

IntelĀ® Performance Counter Monitor (IntelĀ® PCM)

Language:C++License:BSD-3-ClauseStargazers:2745Issues:0Issues:0

fedlearner

A multi-party collaborative machine learning framework

Language:PythonLicense:Apache-2.0Stargazers:892Issues:0Issues:0

Megatron-LM

Ongoing research training transformer models at scale

Language:PythonLicense:NOASSERTIONStargazers:10064Issues:0Issues:0

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Language:PythonLicense:Apache-2.0Stargazers:34843Issues:0Issues:0

attention-is-all-you-need-pytorch

A PyTorch implementation of the Transformer model in "Attention is All You Need".

Language:PythonLicense:MITStargazers:8770Issues:0Issues:0

incubator-mxnet

Lightweight, Portable, Flexible Distributed/Mobile Deep Learning with Dynamic, Mutation-aware Dataflow Dep Scheduler; for Python, R, Julia, Scala, Go, Javascript and more

Language:C++License:Apache-2.0Stargazers:4Issues:0Issues:0

byteps

A high performance and generic framework for distributed DNN training

Language:PythonLicense:NOASSERTIONStargazers:3618Issues:0Issues:0

ps-lite

A lightweight parameter server interface

Language:C++License:Apache-2.0Stargazers:72Issues:0Issues:0

horovod

Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.

Language:PythonLicense:NOASSERTIONStargazers:14169Issues:0Issues:0