Taoshu (TaoLbr1993)

TaoLbr1993

Geek Repo

Company:Alibaba Group

Github PK Tool:Github PK Tool

Taoshu's starred repositories

Language:MATLABLicense:GPL-3.0Stargazers:10356Issues:0Issues:0

KIVI

KIVI: A Tuning-Free Asymmetric 2bit Quantization for KV Cache

Language:PythonLicense:MITStargazers:191Issues:0Issues:0

Pytorch-XNOR-Net

XNOR-Net, with binary gemm and binary conv2d kernels, support both CPU and GPU.

Language:PythonLicense:BSD-3-ClauseStargazers:78Issues:0Issues:0

binary-networks-pytorch

Binarize convolutional neural networks using pytorch :fire:

Language:PythonLicense:BSD-3-ClauseStargazers:130Issues:0Issues:0

kernel_tuner

Kernel Tuner

Language:PythonLicense:Apache-2.0Stargazers:266Issues:0Issues:0

NVIDIA_SGEMM_PRACTICE

Step-by-step optimization of CUDA SGEMM

Language:CudaStargazers:199Issues:0Issues:0

XNOR-Net

ImageNet classification using binary Convolutional Neural Networks

Language:LuaLicense:NOASSERTIONStargazers:857Issues:0Issues:0

BiLLM

(ICML 2024) BiLLM: Pushing the Limit of Post-Training Quantization for LLMs

Language:PythonLicense:MITStargazers:171Issues:0Issues:0

marlin

FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.

Language:PythonLicense:Apache-2.0Stargazers:518Issues:0Issues:0

graphlearn-for-pytorch

A GPU-accelerated graph learning library for PyTorch, facilitating the scaling of GNN training and inference.

Language:PythonLicense:Apache-2.0Stargazers:111Issues:0Issues:0

plot_demo

论文里可以用到的实验图示例

Language:PythonStargazers:184Issues:0Issues:0

neural-compressor

SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime

Language:PythonLicense:Apache-2.0Stargazers:2116Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:4482Issues:0Issues:0

ipex-llm

Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, MiniCPM, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, GraphRAG, DeepSpeed, vLLM, FastChat, Axolotl, etc.

Language:PythonLicense:Apache-2.0Stargazers:6416Issues:0Issues:0

AutoGPTQ

An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.

Language:PythonLicense:MITStargazers:4259Issues:0Issues:0

transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Language:PythonLicense:Apache-2.0Stargazers:130847Issues:0Issues:0

Edge-MoE

Edge-MoE: Memory-Efficient Multi-Task Vision Transformer Architecture with Task-level Sparsity via Mixture-of-Experts

Language:C++Stargazers:80Issues:0Issues:0

LLMSpeculativeSampling

Fast inference from large lauguage models via speculative decoding

Language:PythonLicense:Apache-2.0Stargazers:466Issues:0Issues:0

PygHO

A library for subgraph GNN based on pyg

Language:PythonLicense:MITStargazers:37Issues:0Issues:0
Language:PythonStargazers:256Issues:0Issues:0

FlexGen

Running large language models on a single GPU for throughput-oriented scenarios.

Language:PythonLicense:Apache-2.0Stargazers:9109Issues:0Issues:0

sparsegpt

Code for the ICML 2023 paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot".

Language:PythonLicense:Apache-2.0Stargazers:679Issues:0Issues:0

Megatron-LM

Ongoing research training transformer models at scale

Language:PythonLicense:NOASSERTIONStargazers:9730Issues:0Issues:0

metaseq

Repo for external large-scale work

Language:PythonLicense:MITStargazers:6439Issues:0Issues:0

tuning_playbook

A playbook for systematically maximizing the performance of deep learning models.

License:NOASSERTIONStargazers:26170Issues:0Issues:0

vsc-cec-ide

一个插件,国产化你的VSCode,来源于CEC-IDE,有敏感词检测、防沉迷等功能。

Language:TypeScriptLicense:Apache-2.0Stargazers:777Issues:0Issues:0
Stargazers:226Issues:0Issues:0

llama

Inference code for Llama models

Language:PythonLicense:NOASSERTIONStargazers:55179Issues:0Issues:0

pretrain-gnns

Strategies for Pre-training Graph Neural Networks

Language:PythonLicense:MITStargazers:954Issues:0Issues:0

SUN

Understanding and Extending Subgraph GNNs by Rethinking their Symmetries (NeurIPS 2022 Oral)

Language:PythonLicense:MITStargazers:39Issues:0Issues:0