Yu Zhang (yzhangcs)

yzhangcs

Geek Repo

Company: Soochow University

Location:Shenzhen, Guangdong

Home Page:https://yzhang.site

Twitter:@yzhang_cs

Github PK Tool:Github PK Tool


Organizations
SUDA-LA

Yu Zhang's starred repositories

The-Art-of-Linear-Algebra

Graphic notes on Gilbert Strang's "Linear Algebra for Everyone"

Language:PostScriptLicense:CC0-1.0Stargazers:15317Issues:135Issues:14

LLaMA-Adapter

[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters

Language:PythonLicense:GPL-3.0Stargazers:5585Issues:78Issues:141

gpt-fast

Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.

Language:PythonLicense:BSD-3-ClauseStargazers:5282Issues:61Issues:89

gradient-checkpointing

Make huge neural nets fit in memory

Language:PythonLicense:MITStargazers:2645Issues:81Issues:42

gptq

Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".

Language:PythonLicense:Apache-2.0Stargazers:1765Issues:29Issues:48

S-LoRA

S-LoRA: Serving Thousands of Concurrent LoRA Adapters

Language:PythonLicense:Apache-2.0Stargazers:1579Issues:24Issues:37

segment-anything-fast

A batched offline inference oriented version of segment-anything

Language:PythonLicense:Apache-2.0Stargazers:1136Issues:11Issues:40

Yuan-2.0

Yuan 2.0 Large Language Model

Language:PythonLicense:NOASSERTIONStargazers:664Issues:5Issues:90

linear-attention-transformer

Transformer based on a variant of attention that is linear complexity in respect to sequence length

Language:PythonLicense:MITStargazers:626Issues:12Issues:19

NBCE

Naive Bayes-based Context Extension

AdaLoRA

AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (ICLR 2023).

Language:PythonLicense:MITStargazers:222Issues:1Issues:25

RetNet

Huggingface compatible implementation of RetNet (Retentive Networks, https://arxiv.org/pdf/2307.08621.pdf) including parallel, recurrent, and chunkwise forward.

Language:Jupyter NotebookLicense:MITStargazers:222Issues:5Issues:31

long-llms-learning

A repository sharing the literatures about long-context large language models, including the methodologies and the evaluation benchmarks

Language:Jupyter NotebookStargazers:214Issues:7Issues:2

Mega-pytorch

Implementation of Mega, the Single-head Attention with Multi-headed EMA architecture that currently holds SOTA on Long Range Arena

Language:PythonLicense:MITStargazers:202Issues:8Issues:2

LightSeq

Official repository for LightSeq: Sequence Level Parallelism for Distributed Training of Long Context Transformers

LongQLoRA

LongQLoRA: Extent Context Length of LLMs Efficiently

checkmate

Training neural networks in TensorFlow 2.0 with 5x less memory

Language:PythonLicense:Apache-2.0Stargazers:124Issues:9Issues:57
Language:PythonLicense:MITStargazers:105Issues:4Issues:5

GAU-alpha

基于Gated Attention Unit的Transformer模型(尝鲜版)

grouped-query-attention-pytorch

(Unofficial) PyTorch implementation of grouped-query attention (GQA) from "GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints" (https://arxiv.org/pdf/2305.13245.pdf)

Language:PythonLicense:MITStargazers:81Issues:3Issues:3

minimal-LRU

Non official implementation of the Linear Recurrent Unit (LRU, Orvieto et al. 2023)

Language:PythonLicense:MITStargazers:38Issues:1Issues:3

QKNorm

Code for the paper "Query-Key Normalization for Transformers"

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:32Issues:0Issues:0

make-lstm-great-again

Donald Trump's tweets generator

Language:PythonStargazers:28Issues:2Issues:0

fast_feedforward_computation

Official code for "Accelerating Feedforward Computation via Parallel Nonlinear Equation Solving", ICML 2021

Language:Jupyter NotebookStargazers:24Issues:7Issues:0

mRNN-mLSTM

Code for ICML 2020 paper: Do RNN and LSTM have Long Memory?

Language:PythonLicense:BSD-3-ClauseStargazers:15Issues:0Issues:0
Language:CLicense:MITStargazers:11Issues:0Issues:0