liucc (lcc0504)

lcc0504

Geek Repo

Github PK Tool:Github PK Tool

liucc's starred repositories

alpaca-lora

Instruct-tune LLaMA on consumer hardware

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:18369Issues:155Issues:467

qlora

QLoRA: Efficient Finetuning of Quantized LLMs

Language:Jupyter NotebookLicense:MITStargazers:9675Issues:85Issues:246

GPTQ-for-LLaMa

4 bits quantization of LLaMA using GPTQ

Language:PythonLicense:Apache-2.0Stargazers:2945Issues:42Issues:216

neural-compressor

SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime

Language:PythonLicense:Apache-2.0Stargazers:2065Issues:34Issues:188

gptq

Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".

Language:PythonLicense:Apache-2.0Stargazers:1779Issues:29Issues:48

awesome-model-quantization

A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.

smoothquant

[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models

Language:PythonLicense:MITStargazers:1094Issues:19Issues:81

DiffuSeq

[ICLR'23] DiffuSeq: Sequence to Sequence Text Generation with Diffusion Models

Language:PythonLicense:MITStargazers:698Issues:26Issues:78

OmniQuant

[ICLR2024 spotlight] OmniQuant is a simple and powerful quantization technique for LLMs.

Language:PythonLicense:MITStargazers:613Issues:16Issues:69

SqueezeLLM

[ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization

Language:PythonLicense:MITStargazers:593Issues:17Issues:25
Language:PythonLicense:Apache-2.0Stargazers:513Issues:22Issues:21

MS-AMP

Microsoft Automatic Mixed Precision Library

Language:PythonLicense:MITStargazers:479Issues:11Issues:57

QuIP

Code for paper: "QuIP: 2-Bit Quantization of Large Language Models With Guarantees"

MambaIR

A simple baseline for image restoration with state-space model.

Language:PythonLicense:Apache-2.0Stargazers:297Issues:5Issues:34

q-diffusion

[ICCV 2023] Q-Diffusion: Quantizing Diffusion Models.

Language:PythonLicense:MITStargazers:283Issues:17Issues:36

Atom

[MLSys'24] Atom: Low-bit Quantization for Efficient and Accurate LLM Serving

BlackMamba

Code repository for Black Mamba

language-model-arithmetic

Controlled Text Generation via Language Model Arithmetic

Language:PythonLicense:MITStargazers:182Issues:8Issues:7

RPTQ4LLM

Reorder-based post-training quantization for large language model

Language:PythonLicense:MITStargazers:176Issues:7Issues:12

LLM-FP4

The official implementation of the EMNLP 2023 paper LLM-FP4

Language:PythonLicense:MITStargazers:145Issues:5Issues:9

PTQ4DM

Implementation of Post-training Quantization on Diffusion Models (CVPR 2023)

PTQD

The official implementation of PTQD: Accurate Post-Training Quantization for Diffusion Models

Language:Jupyter NotebookStargazers:81Issues:5Issues:16

disentangle-semantics-syntax

Code for "A Multi-Task Approach for Disentangling Syntax and Semantics in Sentence Representations" (NAACL 2019)

Awesome-Mamba-in-Low-Level-Vision

A paper list of recent mamba efforts for low-level vision.

License:MITStargazers:56Issues:0Issues:0

QuantSR

[NeurIPS 2023 Spotlight] This project is the official implementation of our accepted NeurIPS 2023 (spotlight) paper QuantSR: Accurate Low-bit Quantization for Efficient Image Super-Resolution.

Language:PythonLicense:Apache-2.0Stargazers:37Issues:3Issues:3

DDTB

Pytorch implementation of our paper accepted by ECCV2022 -- Dynamic Dual Trainable Bounds for Ultra-low Precision Super-Resolution Networks

Discriminator-Cooperative-Unlikelihood-Prompt-Tuning

The code implementation of the EMNLP2022 paper: DisCup: Discriminator Cooperative Unlikelihood Prompt-tuning for Controllable Text Generation

Residual_Memory_Transformer

This repository contains code, data, checkpoints, and training and evaluation instructions for the paper: Controllable Text Generation with Residual Memory Transformer