Yi Liu (yiliu30)

yiliu30

Geek Repo

Company:AI Frameworks Engineer @Intel

Location:SH

Github PK Tool:Github PK Tool

Yi Liu's repositories

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

neural-compressor

Intel® Neural Compressor (formerly known as Intel® Low Precision Optimization Tool), targeting to provide unified APIs for network compression technologies, such as low precision quantization, sparsity, pruning, knowledge distillation, across different deep learning frameworks to pursue optimal inference performance.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

oneDNN

oneAPI Deep Neural Network Library (oneDNN)

Language:C++License:Apache-2.0Stargazers:0Issues:0Issues:0

pytorch

Tensors and Dynamic neural networks in Python with strong GPU acceleration

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

torchao-fork

The torchao repository contains api's and workflows for quantization and pruning gpu models.

Language:PythonLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0

accelerate

🚀 A simple way to train and use PyTorch models with multi-GPU, TPU, mixed-precision

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

ai-pr-reviewer

AI-based Pull Request Summarizer and Reviewer with Chat Capabilities.

Language:TypeScriptLicense:MITStargazers:0Issues:0Issues:0

auto-awq-fork

AutoAWQ implements the AWQ algorithm for 4-bit quantization with a 2x speedup during inference. Documentation:

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

auto-round

SOTA Weight-only Quantization Algorithm for LLMs

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

CodeXGLUE

CodeXGLUE

Language:C#License:MITStargazers:0Issues:0Issues:0

gemma.cpp

lightweight, standalone C++ inference engine for Google's Gemma models.

License:Apache-2.0Stargazers:0Issues:0Issues:0

gpt-fast

Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.

Language:PythonLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0

hqq

Official implementation of Half-Quadratic Quantization (HQQ)

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

intel-extension-for-transformers

⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡

Language:C++License:Apache-2.0Stargazers:0Issues:0Issues:0

ipex-llm

Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, Axolotl, etc.

License:Apache-2.0Stargazers:0Issues:0Issues:0

marlin

FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

mpi-operator

Kubernetes Operator for MPI-based applications (distributed training, HPC, etc.)

License:Apache-2.0Stargazers:0Issues:0Issues:0

nn-zero-to-hero

Neural Networks: Zero to Hero

Language:Jupyter NotebookLicense:MITStargazers:0Issues:0Issues:0

optimum-habana

Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

optimum-intel

🤗 Optimum Intel: Accelerate inference with Intel optimization tools

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:1Issues:1
Language:Jupyter NotebookStargazers:0Issues:0Issues:0

tgi

Large Language Model Text Generation Inference

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Torch-Fx-Graph-Visualizer

Visualizer for neural network, deep learning and machine learning models

Language:JavaScriptLicense:MITStargazers:0Issues:0Issues:0

training-operator

Training operators on Kubernetes.

Language:GoLicense:Apache-2.0Stargazers:0Issues:0Issues:0

transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

tutorials

PyTorch tutorials.

Language:PythonLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0

xTuring

Easily build, customize and control your own LLMs

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Stargazers:0Issues:0Issues:0
Language:HTMLStargazers:0Issues:0Issues:0