Yu Zhang's starred repositories
The-Art-of-Linear-Algebra
Graphic notes on Gilbert Strang's "Linear Algebra for Everyone"
LLaMA-Adapter
[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
gradient-checkpointing
Make huge neural nets fit in memory
segment-anything-fast
A batched offline inference oriented version of segment-anything
linear-attention-transformer
Transformer based on a variant of attention that is linear complexity in respect to sequence length
long-llms-learning
A repository sharing the literatures about long-context large language models, including the methodologies and the evaluation benchmarks
Mega-pytorch
Implementation of Mega, the Single-head Attention with Multi-headed EMA architecture that currently holds SOTA on Long Range Arena
grouped-query-attention-pytorch
(Unofficial) PyTorch implementation of grouped-query attention (GQA) from "GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints" (https://arxiv.org/pdf/2305.13245.pdf)
minimal-LRU
Non official implementation of the Linear Recurrent Unit (LRU, Orvieto et al. 2023)
make-lstm-great-again
Donald Trump's tweets generator
fast_feedforward_computation
Official code for "Accelerating Feedforward Computation via Parallel Nonlinear Equation Solving", ICML 2021
mRNN-mLSTM
Code for ICML 2020 paper: Do RNN and LSTM have Long Memory?