Zhiqing Sun (Edward-Sun)

Edward-Sun

Geek Repo

Company:Carnegie Mellon University

Location:Pittsburgh

Home Page:zhiqingsun.com

Twitter:@EdwardSun0909

Github PK Tool:Github PK Tool

Zhiqing Sun's starred repositories

mlc-llm

Universal LLM Deployment Engine with ML Compilation

Language:PythonLicense:Apache-2.0Stargazers:17361Issues:165Issues:1112

llama-recipes

Scripts for fine-tuning Llama2 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization & question answering. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment.Demo apps to showcase Llama2 for WhatsApp & Messenger

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:7850Issues:68Issues:227

DeepSeek-Coder

DeepSeek Coder: Let the Code Write Itself

Language:PythonLicense:MITStargazers:5691Issues:63Issues:142

gpt-fast

Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.

Language:PythonLicense:BSD-3-ClauseStargazers:5246Issues:60Issues:87

AutoGPTQ

An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.

Language:PythonLicense:MITStargazers:3940Issues:34Issues:421
Language:PythonLicense:Apache-2.0Stargazers:3754Issues:51Issues:103

Eureka

Official Repository for "Eureka: Human-Level Reward Design via Coding Large Language Models" (ICLR 2024)

Language:Jupyter NotebookLicense:MITStargazers:2687Issues:25Issues:33

FlexFlow

FlexFlow Serve: Low-Latency, High-Performance LLM Serving

Language:C++License:Apache-2.0Stargazers:1556Issues:32Issues:603

OpenRLHF

An Easy-to-use, Scalable and High-performance RLHF Framework (Support 70B+ full tuning & LoRA & Mixtral & KTO)

Language:PythonLicense:Apache-2.0Stargazers:1453Issues:20Issues:146

LLMTest_NeedleInAHaystack

Doing simple retrieval from LLM models at various context lengths to measure accuracy

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:1163Issues:12Issues:25

hallucination-leaderboard

Leaderboard Comparing LLM Performance at Producing Hallucinations when Summarizing Short Documents

dolma

Data and tools for generating and inspecting OLMo pre-training data.

Language:PythonLicense:Apache-2.0Stargazers:814Issues:17Issues:61

factool

FacTool: Factuality Detection in Generative AI

Language:PythonLicense:Apache-2.0Stargazers:772Issues:10Issues:28

PiPPy

Pipeline Parallelism for PyTorch

Language:PythonLicense:BSD-3-ClauseStargazers:646Issues:36Issues:247

representation-engineering

Representation Engineering: A Top-Down Approach to AI Transparency

Language:Jupyter NotebookLicense:MITStargazers:592Issues:29Issues:35

gpt_paper_assistant

GPT4 based personalized ArXiv paper assistant bot

Language:PythonLicense:Apache-2.0Stargazers:431Issues:6Issues:10

weatherbench2

A benchmark for the next generation of data-driven global weather models.

Language:PythonLicense:Apache-2.0Stargazers:334Issues:8Issues:34

JudgeLM

An open-sourced LLM judge for evaluating LLM-generated answers.

Language:PythonLicense:Apache-2.0Stargazers:268Issues:7Issues:15

UltraFeedback

A large-scale, fine-grained, diverse preference dataset (and models).

Language:PythonLicense:MITStargazers:268Issues:10Issues:13

LLaVA-RLHF

Aligning LMMs with Factually Augmented RLHF

Language:PythonLicense:GPL-3.0Stargazers:250Issues:8Issues:30

LRV-Instruction

[ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning

Language:PythonLicense:BSD-3-ClauseStargazers:225Issues:11Issues:22

ModuleFormer

ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward experts. We released a collection of ModuleFormer-based Language Models (MoLM) ranging in scale from 4 billion to 8 billion parameters.

Language:PythonLicense:Apache-2.0Stargazers:215Issues:11Issues:5

Copyisallyouneed

[ICLR 2023] Codebase for Copy-Generator model, including an implementation of kNN-LM

Language:PythonLicense:MITStargazers:178Issues:4Issues:14

SALMON

Self-Alignment with Principle-Following Reward Models

Language:PythonLicense:GPL-3.0Stargazers:126Issues:5Issues:1
Language:PythonLicense:MITStargazers:92Issues:2Issues:5

triton-autodiff

Experiment of using Tangent to autodiff triton

Language:PythonLicense:MITStargazers:66Issues:6Issues:0

miniF2F

An updated version of miniF2F with lots of fixes and informal statements / solutions.

Language:Objective-C++License:MITStargazers:51Issues:9Issues:8

CL-LNS

Code repo for ICML'23 Searching Large Neighborhoods for Integer Linear Programs with Contrastive Learning

Language:PythonLicense:NOASSERTIONStargazers:32Issues:3Issues:2

Accelerating-Diffusion-based-Combinatorial-Optimization-Solvers-by-Progressive-Distillation

Code for Accelerating Diffusion-based Combinatorial Optimization Solvers by Progressive Distillation

Stargazers:5Issues:0Issues:0