Yu Zhang (yzhangcs)

yzhangcs

Geek Repo

Company: Soochow University

Location:Shanghai

Home Page:https://yzhang.site

Twitter:@yzhang_cs

Github PK Tool:Github PK Tool


Organizations
SUDA-LA

Yu Zhang's starred repositories

LMFlow

An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.

Language:PythonLicense:Apache-2.0Stargazers:8187Issues:73Issues:402

torchtune

A Native-PyTorch Library for LLM Fine-tuning

Language:PythonLicense:BSD-3-ClauseStargazers:3824Issues:43Issues:450

LLMTest_NeedleInAHaystack

Doing simple retrieval from LLM models at various context lengths to measure accuracy

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:1401Issues:15Issues:25

nanotron

Minimalistic large language model 3D-parallelism training

Language:PythonLicense:Apache-2.0Stargazers:1049Issues:42Issues:72

JetMoE

Reaching LLaMA2 Performance with 0.1M Dollars

Language:PythonLicense:Apache-2.0Stargazers:953Issues:8Issues:9

tensor_parallel

Automatically split your PyTorch models on multiple GPUs for training & inference

Language:PythonLicense:MITStargazers:612Issues:8Issues:66

ByteTransformer

optimized BERT transformer inference on NVIDIA GPU. https://arxiv.org/abs/2210.03052

Language:C++License:Apache-2.0Stargazers:448Issues:10Issues:10

InfiniTransformer

Unofficial PyTorch/🤗Transformers(Gemma/Llama3) implementation of Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention

Language:PythonLicense:MITStargazers:327Issues:8Issues:24

long-context-attention

Sequence Parallel Attention for Long Context LLM Model Training and Inference

Language:PythonLicense:Apache-2.0Stargazers:267Issues:4Issues:14

Rewrite-the-Stars

[CVPR 2024] Rewrite the Stars

Language:PythonLicense:Apache-2.0Stargazers:239Issues:2Issues:17

minicons

Utility for behavioral and representational analyses of Language Models

Language:PythonLicense:MITStargazers:113Issues:6Issues:16

DiJiang

[ICML'24 Oral] The official code of "DiJiang: Efficient Large Language Models through Compact Kernelization", a novel DCT-based linear attention mechanism.

mad-lab

A MAD laboratory to improve AI architecture designs 🧪

Language:PythonLicense:MITStargazers:82Issues:1Issues:2

LLM-Inheritune

This is the official repository for Inheritune.

LASP

Linear Attention Sequence Parallelism (LASP)

Language:PythonLicense:MITStargazers:61Issues:2Issues:0

LinearAttentionArena

Here we will test various linear attention designs.

Language:PythonLicense:Apache-2.0Stargazers:53Issues:8Issues:0

infini-mini-transformer

This is a personal reimplementation of Google's Infini-transformer, utilizing a small 2b model. The project includes both model and training code.

gpt-accelera

Simple and efficient pytorch-native transformer training and inference (batched)

Language:PythonLicense:BSD-3-ClauseStargazers:50Issues:3Issues:0
Language:PythonStargazers:46Issues:6Issues:0

HGRN2

HGRN2: Gated Linear RNNs with State Expansion

Language:PythonStargazers:42Issues:1Issues:0

rwkv

RWKV model implementation

Language:PythonLicense:MITStargazers:38Issues:2Issues:0

GORU-tensorflow

Gated Orthogonal Recurrent Unit implementation in tensorflow

Language:PythonLicense:MITStargazers:35Issues:5Issues:2

rnn-icrag

Official repository of paper "RNNs Are Not Transformers (Yet): The Key Bottleneck on In-context Retrieval"

Language:PythonStargazers:24Issues:2Issues:0

heinsen_attention

Reference implementation of "Softmax Attention with Constant Cost per Token" (Heinsen, 2024)

Language:PythonLicense:MITStargazers:22Issues:3Issues:1
Language:PythonLicense:Apache-2.0Stargazers:18Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:11Issues:0Issues:0
Language:PythonLicense:MITStargazers:2Issues:1Issues:0