Jue WANG's repositories
ChatGPT-Next-Web
A cross-platform ChatGPT/Gemini UI (Web / PWA / Linux / Win / MacOS). 一键拥有你自己的跨平台 ChatGPT/Gemini 应用。
lm-evaluation-harness
A framework for few-shot evaluation of autoregressive language models.
helm
Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110).
LorrinWWW.github.io
Jue's blog
FLASK
[ICLR 2024 Spotlight] FLASK: Fine-grained Language Model Evaluation based on Alignment Skill Sets
FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and FastChat-T5.
LLM-Pruner
[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support LLaMA, Llama-2, BLOOM, Vicuna, Baichuan, etc.
lm-format-enforcer
Enforce the output format (JSON Schema, Regex etc) of a language model
gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
advertorch
A Toolbox for Adversarial Robustness Research
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
fmengine-torch
FMEngine [PyTorch version]
scaled-rope
Lora
json-stream
Simple streaming JSON parser and encoder.
natural-instructions
Expanding natural instructions
RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
DeeperSpeed
DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.
Open-Instruction-Generalist
Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks
open_clip
An open source implementation of CLIP.
icetk
A unified tokenization tool for Images, Chinese and English.
GLM-130B
GLM-130B: An Open Bilingual Pre-Trained Model
min-dalle
min(DALL·E) is a fast, minimal port of DALL·E Mega to PyTorch