Fangkai Jiao (SparkJiao)

SparkJiao

Geek Repo

Company:NTU-NLP & I2R, A*STAR, Singapore

Location:Sinagpore

Home Page:jiaofangkai.com

Github PK Tool:Github PK Tool

Fangkai Jiao's starred repositories

dspy

DSPy: The framework for programming—not prompting—foundation models

Language:PythonLicense:MITStargazers:11803Issues:114Issues:474

gpt-fast

Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.

Language:PythonLicense:BSD-3-ClauseStargazers:5215Issues:59Issues:86

gemma_pytorch

The official PyTorch implementation of Google's Gemma models

Language:PythonLicense:Apache-2.0Stargazers:5070Issues:39Issues:34

OLMo

Modeling, training, eval, and inference code for OLMo

Language:PythonLicense:Apache-2.0Stargazers:4079Issues:41Issues:153

mergekit

Tools for merging pretrained large language models.

Language:PythonLicense:LGPL-3.0Stargazers:3726Issues:45Issues:228

Qwen1.5

Qwen1.5 is the improved version of Qwen, the large language model series developed by Qwen team, Alibaba Cloud.

Pearl

A Production-ready Reinforcement Learning AI Agent Library brought by the Applied Reinforcement Learning team at Meta.

Language:Jupyter NotebookLicense:MITStargazers:2396Issues:29Issues:50

lightllm

LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.

Language:PythonLicense:Apache-2.0Stargazers:1899Issues:19Issues:161
Language:PythonLicense:Apache-2.0Stargazers:1032Issues:13Issues:78

dlrover

DLRover: An Automatic Distributed Deep Learning System

Language:PythonLicense:NOASSERTIONStargazers:990Issues:49Issues:210

nanotron

Minimalistic large language model 3D-parallelism training

Language:PythonLicense:Apache-2.0Stargazers:824Issues:40Issues:54
Language:PythonLicense:Apache-2.0Stargazers:765Issues:12Issues:34

llama-moe

⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training

Language:PythonLicense:Apache-2.0Stargazers:747Issues:8Issues:18

alpaca_farm

A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.

Language:PythonLicense:Apache-2.0Stargazers:721Issues:8Issues:41

EAGLE

[ICML'24] EAGLE: Speculative Sampling Requires Rethinking Feature Uncertainty

Language:PythonLicense:Apache-2.0Stargazers:524Issues:10Issues:61

flash-linear-attention

Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton

Language:PythonLicense:MITStargazers:520Issues:18Issues:14

lumos

Code and data for "Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs"

Language:PythonLicense:MITStargazers:416Issues:9Issues:4

CoLLiE

Collaborative Training of Large Language Models in an Efficient Way

Language:PythonLicense:Apache-2.0Stargazers:390Issues:10Issues:69

Awesome-Reasoning-Foundation-Models

✨✨Latest Papers and Benchmarks in Reasoning with Foundation Models

ChunkLlama

[ICML'24] Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"

Language:PythonLicense:Apache-2.0Stargazers:222Issues:8Issues:12

zero-bubble-pipeline-parallelism

Zero Bubble Pipeline Parallelism

Language:PythonLicense:NOASSERTIONStargazers:196Issues:5Issues:12

HallusionBench

[CVPR'24] HallusionBench: You See What You Think? Or You Think What You See? An Image-Context Reasoning Benchmark Challenging for GPT-4V(ision), LLaVA-1.5, and Other Multi-modality Models

Language:PythonLicense:BSD-3-ClauseStargazers:183Issues:4Issues:10

Q-Instruct

②[CVPR 2024] Low-level visual instruction tuning, with a 200K dataset and a model zoo for fine-tuned checkpoints.

Language:PythonLicense:MITStargazers:162Issues:2Issues:22
Language:PythonLicense:Apache-2.0Stargazers:68Issues:2Issues:2

RLCD

Reproduction of "RLCD Reinforcement Learning from Contrast Distillation for Language Model Alignment

Language:PythonLicense:MITStargazers:54Issues:5Issues:3

SimulateBench

GPT as Human

Language:PythonStargazers:18Issues:0Issues:0

SeaEval

NAACL 2024: SeaEval for Multilingual Foundation Models: From Cross-Lingual Alignment to Cultural Reasoning

Language:PythonLicense:NOASSERTIONStargazers:12Issues:0Issues:1

dpo-trajectory-reasoning

Source code for the paper "Learning Planning-based Reasoning with Trajectory Collection and Process Rewards Synthesizing".

Language:PythonStargazers:11Issues:2Issues:0

RLMEC

The official repository of "Improving Large Language Models via Fine-grained Reinforcement Learning with Minimum Editing Constraint"

Language:PythonStargazers:6Issues:0Issues:0