J.L's starred repositories

Language:PythonStargazers:5Issues:0Issues:0

PiSSA

PiSSA: Principal Singular Values and Singular Vectors Adaptation of Large Language Models(NeurIPS 2024 Spotlight)

Language:Jupyter NotebookStargazers:248Issues:0Issues:0

efficient-kan

An efficient pure-PyTorch implementation of Kolmogorov-Arnold Network (KAN).

Language:PythonLicense:MITStargazers:3923Issues:0Issues:0

persona-hub

Official repo for the paper "Scaling Synthetic Data Creation with 1,000,000,000 Personas"

Language:PythonStargazers:805Issues:0Issues:0

RWKV_LM_EXT

This project is to extend RWKV LM's capabilities including sequence classification/embedding/peft/cross encoder/bi encoder/multi modalities, etc.

Language:PythonStargazers:7Issues:0Issues:0

uncheatable_eval

Evaluating LLMs with Dynamic Data

Language:Jupyter NotebookLicense:MITStargazers:66Issues:0Issues:0

rwkv-tokenizer

A fast RWKV Tokenizer written in Rust

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:35Issues:0Issues:0

Megatron-LM

Ongoing research training transformer models at scale

License:NOASSERTIONStargazers:2Issues:0Issues:0

ollama

Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.

Language:GoLicense:MITStargazers:91641Issues:0Issues:0

RWKV-AlignBench

RWKV模型中文对齐评测结果

Language:Jupyter NotebookStargazers:1Issues:0Issues:0
Language:PythonStargazers:2Issues:0Issues:0
Language:PythonStargazers:76Issues:0Issues:0

flash-linear-attention

Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton

Language:PythonLicense:MITStargazers:1235Issues:0Issues:0
Language:PythonStargazers:9Issues:0Issues:0

Mat2Stencil

A Modular Matrix-Based DSL for Explicit and Implicit Matrix-Free PDE Solvers on Structured Grid.

Stargazers:2Issues:0Issues:0
Language:PythonStargazers:5Issues:0Issues:0

RWKV-infctx-trainer

RWKV infctx trainer, for training arbitary context sizes, to 10k and beyond!

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:132Issues:0Issues:0

VisualRWKV

VisualRWKV is the visual-enhanced version of the RWKV language model, enabling RWKV to handle various visual tasks.

Language:PythonLicense:Apache-2.0Stargazers:173Issues:0Issues:0

RWKV-CUDA

The CUDA version of the RWKV language model ( https://github.com/BlinkDL/RWKV-LM )

Language:CudaStargazers:4Issues:0Issues:0

RWKV-LM

RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.

Language:PythonLicense:Apache-2.0Stargazers:12466Issues:0Issues:0

ChatRWKV

ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.

Language:PythonLicense:Apache-2.0Stargazers:9391Issues:0Issues:0

speech_dataset

The dataset of Speech Recognition

License:Apache-2.0Stargazers:383Issues:0Issues:0
Language:PythonStargazers:5Issues:0Issues:0
Language:PythonStargazers:10Issues:0Issues:0