Jue WANG (LorrinWWW)

LorrinWWW

Geek Repo

Company:Zhejiang University

Location:San Francisco

Home Page:https://juewang.me/about/

Github PK Tool:Github PK Tool

Jue WANG's repositories

SkipBERT

Code associated with the paper **SkipBERT: Efficient Inference with Shallow Layer Skipping**, at ACL 2022

Language:PythonLicense:Apache-2.0Stargazers:15Issues:5Issues:1

Snapshot

Code associated with the paper Effective Continual Learning for Text Classification with Lightweight Snapshots, at AAAI 2023

Language:PythonStargazers:1Issues:2Issues:0

DeeperSpeed

DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

GLM-130B

GLM-130B: An Open Bilingual Pre-Trained Model

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

advertorch

A Toolbox for Adversarial Robustness Research

Language:Jupyter NotebookLicense:LGPL-3.0Stargazers:0Issues:0Issues:0

ChatGPT-Next-Web

A cross-platform ChatGPT/Gemini UI (Web / PWA / Linux / Win / MacOS). 一键拥有你自己的跨平台 ChatGPT/Gemini 应用。

Language:TypeScriptLicense:MITStargazers:0Issues:0Issues:0
Language:DockerfileStargazers:0Issues:0Issues:0

FastChat

An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and FastChat-T5.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

FLASK

[ICLR 2024 Spotlight] FLASK: Fine-grained Language Model Evaluation based on Alignment Skill Sets

Language:PythonStargazers:0Issues:0Issues:0

fmengine-torch

FMEngine [PyTorch version]

Language:PythonStargazers:0Issues:0Issues:0

gpt-neox

An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:1Issues:0

helm

Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110).

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

icetk

A unified tokenization tool for Images, Chinese and English.

Language:PythonStargazers:0Issues:0Issues:0

json-stream

Simple streaming JSON parser and encoder.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

LLM-Pruner

[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support LLaMA, Llama-2, BLOOM, Vicuna, Baichuan, etc.

License:Apache-2.0Stargazers:0Issues:0Issues:0

lm-evaluation-harness

A framework for few-shot evaluation of autoregressive language models.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

lm-format-enforcer

Enforce the output format (JSON Schema, Regex etc) of a language model

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

min-dalle

min(DALL·E) is a fast, minimal port of DALL·E Mega to PyTorch

Language:PythonLicense:MITStargazers:0Issues:1Issues:0
Language:PythonStargazers:0Issues:1Issues:0

natural-instructions

Expanding natural instructions

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Open-Instruction-Generalist

Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

open_clip

An open source implementation of CLIP.

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:0Issues:0

RedPajama-Data

The RedPajama-Data repository contains code for preparing large datasets for training large language models.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:0Issues:0
Stargazers:0Issues:1Issues:0

vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0