Leyang Xue's repositories

drunkcoding.github.io

everything on distributed file system and cloud storage

Language:HTMLLicense:MITStargazers:2Issues:1Issues:0
Language:PythonLicense:MITStargazers:1Issues:2Issues:0

model-inference

utilities and tests for model inference

Language:PythonLicense:MITStargazers:1Issues:2Issues:0

alpa

Training and serving large-scale neural networks with auto parallelization.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
License:MITStargazers:0Issues:1Issues:0

cheetah-fastclick

FastClick with the Cheetah elements

Language:C++License:NOASSERTIONStargazers:0Issues:1Issues:0
Language:RLicense:GPL-3.0Stargazers:0Issues:2Issues:0

core

The core library and APIs implementing the Triton Inference Server.

Language:C++License:BSD-3-ClauseStargazers:0Issues:1Issues:0

CS411-Database-System

Project for database system -- an interactive website

Language:PythonLicense:GPL-3.0Stargazers:0Issues:3Issues:0

DeeperSpeed

DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.

Language:PythonLicense:MITStargazers:0Issues:1Issues:0
License:Apache-2.0Stargazers:0Issues:2Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:2Issues:0

eudyptula

linux kernel challenge

Language:CLicense:GPL-3.0Stargazers:0Issues:2Issues:0

falcon

FALCON - Fast Analysis of LTE Control channels

Language:C++License:AGPL-3.0Stargazers:0Issues:1Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:1Issues:0

gpt-neox

An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:1Issues:0

jiant

jiant is an nlp toolkit

Language:PythonLicense:MITStargazers:0Issues:1Issues:0
Language:GoLicense:MITStargazers:0Issues:2Issues:0

mixtral-offloading

Run Mixtral-8x7B models in Colab or consumer desktops

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

model-finetune

finetune pre-trained models

Language:PythonLicense:MITStargazers:0Issues:2Issues:0

onnxruntime_backend

The Triton backend for the ONNX Runtime.

Language:C++License:BSD-3-ClauseStargazers:0Issues:1Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

power-meter

A software power measurement tool for both CPU and GPU using vendor provided API

Language:C++License:Apache-2.0Stargazers:0Issues:2Issues:0

pytorch_backend

The Triton backend for the PyTorch TorchScript models.

Language:C++License:BSD-3-ClauseStargazers:0Issues:1Issues:0

server

The Triton Inference Server provides an optimized cloud and edge inferencing solution.

Language:PythonLicense:BSD-3-ClauseStargazers:0Issues:1Issues:0

simple-shell

Simple functioning shell implemented in C

Language:CStargazers:0Issues:2Issues:0
Language:PythonLicense:MITStargazers:0Issues:2Issues:0
Language:PythonStargazers:0Issues:2Issues:0

transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

License:Apache-2.0Stargazers:0Issues:0Issues:0

wasmint

Library for interpreting / debugging wasm code

Language:C++License:Apache-2.0Stargazers:0Issues:1Issues:0