Shudong Liu (sudanl)

sudanl

Geek Repo

Company:University of Macau

Location:Macau SAR, China

Home Page:sudanl.github.io

Twitter:@shudong_liu

Github PK Tool:Github PK Tool


Organizations
microsoft
MicrosoftCopilot
NLP2CT

Shudong Liu's starred repositories

llm-course

Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:34928Issues:364Issues:65

llama3

The official Meta Llama 3 GitHub site

Language:PythonLicense:NOASSERTIONStargazers:24971Issues:206Issues:212

vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

Language:PythonLicense:Apache-2.0Stargazers:23924Issues:217Issues:3685

WizardLM

LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMath

lmdeploy

LMDeploy is a toolkit for compressing, deploying, and serving LLMs.

Language:PythonLicense:Apache-2.0Stargazers:3595Issues:33Issues:1157

CampusShame

互联网仍有记忆!那些曾经在校招过程中毁过口头offer、意向书、三方的公司!纵然人微言轻,也想尽绵薄之力!

alpaca_eval

An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:1353Issues:7Issues:133

DeepSeek-LLM

DeepSeek LLM: Let there be answers

Language:MakefileLicense:MITStargazers:1349Issues:23Issues:32

textgrad

TextGrad: Automatic ''Differentiation'' via Text -- using large language models to backpropagate textual gradients.

Language:PythonLicense:MITStargazers:1263Issues:19Issues:48

DeepSeek-MoE

DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models

Language:PythonLicense:MITStargazers:934Issues:15Issues:35

llama-moe

⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training

Language:PythonLicense:Apache-2.0Stargazers:812Issues:7Issues:18

DeepSeek-Math

DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models

Language:PythonLicense:MITStargazers:731Issues:13Issues:24

Awesome-Knowledge-Distillation-of-LLMs

This repository collects papers for "A Survey on Knowledge Distillation of Large Language Models". We break down KD into Knowledge Elicitation and Distillation Algorithms, and explore the Skill & Vertical Distillation of LLMs.

deita

Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]

Language:PythonLicense:Apache-2.0Stargazers:429Issues:6Issues:24

dl4math

Resources of deep learning for mathematical reasoning (DL4MATH).

magpie

Official repository for "Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing". Your efficient and high-quality synthetic data generation pipeline!

Language:PythonLicense:MITStargazers:273Issues:6Issues:13

Token-level-Direct-Preference-Optimization

Reference implementation for Token-level Direct Preference Optimization(TDPO)

Language:PythonLicense:Apache-2.0Stargazers:76Issues:1Issues:3

easy-to-hard-generalization

Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"

Language:PythonLicense:Apache-2.0Stargazers:44Issues:6Issues:0

CalibratedMath

Teaching Models to Express Their Uncertainty in Words

awesome-cultural-nlp

Resources for cultural NLP research

awesome-instruction-selector

Paper list and datasets for the paper: A Survey on Data Selection for LLM Instruction Tuning

Stargazers:22Issues:0Issues:0

MathCheck

Is Your Model Really A Good Math Reasoner? Evaluating Mathematical Reasoning with Checklist

Language:PythonStargazers:18Issues:0Issues:0

privacy-preserving-prompt

Privacy-Preserving Prompt Tuning for Large Language Model

Language:PythonStargazers:9Issues:0Issues:0

TempoSum

Can LMs Generalize to Future Data? An Empirical Analysis on Text Summarization