PENG Bo's repositories
RWKV-LM
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.
minGPT-tuned
A *tuned* minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
modded-nanogpt-rwkv
RWKV-7: Surpassing GPT
RWKV-v2-RNN-Pile
RWKV-v2-RNN trained on the Pile. See https://github.com/BlinkDL/RWKV-LM for details.
LinearAttentionArena
Here we will test various linear attention designs.
SmallInitEmb
LayerNorm(SmallInit(Embedding)) in a Transformer to improve convergence
WorldModel
Let us make Psychohistory (as in Asimov) a reality, and accessible to everyone. Useful for LLM grounding and games / fiction / business / finance / governance, and can align agents with human too.
LM-Trick-Questions
Here we collect trick questions and failed tasks for open source LLMs to improve them.
BlinkColorTheme
A colorful theme for HTML+JS+CSS.
Model_Leaderboard
Leaderboard of AI models.
BasisLang.com
BasisLang.com
ProjectIvory
Project Ivory is a simple forum written a few years ago.