GUO-QING JIANG's repositories
Auto-GPT
An experimental open-source attempt to make GPT-4 fully autonomous.
awesome-chatgpt-prompts
This repo includes ChatGPT prompt curation to use ChatGPT better.
Awesome-Deep-Neural-Network-Compression
Summary, Code for Deep Neural Network Quantization
BELLE
BELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)
ChatGLM-6B
ChatGLM-6B:开源双语对话语言模型
Chinese-LLaMA-Alpaca
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
CLIP
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
ColossalAI
Making big AI models cheaper, easier, and more scalable
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
DeepSpeedExamples
Example models using DeepSpeed
fairseq
Facebook AI Research Sequence-to-Sequence Toolkit written in Python.
FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and FastChat-T5.
langchain
⚡ Building applications with LLMs through composability ⚡
LAVIS
LAVIS - A One-stop Library for Language-Vision Intelligence
lightseq
LightSeq: A High Performance Library for Sequence Processing and Generation
llama
Inference code for LLaMA models
llama.cpp
LLM inference in C/C++
Medusa
Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Heads
Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
MS-AMP
Microsoft Automatic Mixed Precision Library
NeMo
NeMo: a framework for generative AI
openai-cookbook
Examples and guides for using the OpenAI API
peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
rtp-llm
RTP-LLM: Alibaba's high-performance LLM inference engine for diverse applications.
stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
stylegan-xl
[SIGGRAPH'22] StyleGAN-XL: Scaling StyleGAN to Large Diverse Datasets
stylegan2-ada-pytorch
StyleGAN2-ADA - Official PyTorch implementation
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
trlx
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)