YP's starred repositories

LLM-PowerHouse-A-Curated-Guide-for-Large-Language-Models-with-Custom-Training-and-Inferencing

LLM-PowerHouse: Unleash LLMs' potential through curated tutorials, best practices, and ready-to-use code for custom training and inferencing.

Language:Jupyter NotebookLicense:MITStargazers:479Issues:0Issues:0

ThunderKittens

Tile primitives for speedy kernels

Language:CudaLicense:MITStargazers:972Issues:0Issues:0

tiny-universe

《大模型白盒子构建指南》:一个全手搓的Tiny-Universe

Language:PythonStargazers:249Issues:0Issues:0

ccml

simple autodiff library

Language:CStargazers:60Issues:0Issues:0

tiny-gpu

A minimal GPU design in Verilog to learn how GPUs work from the ground up

Language:SystemVerilogStargazers:5933Issues:0Issues:0

torchtitan

A native PyTorch Library for large model training

Language:PythonLicense:BSD-3-ClauseStargazers:1147Issues:0Issues:0

llm-reasoners

A library for advanced large language model reasoning

Language:PythonLicense:Apache-2.0Stargazers:889Issues:0Issues:0

Shiro

📜 A minimalist personal website embodying the purity of paper and freshness of snow.

Language:TypeScriptLicense:NOASSERTIONStargazers:2791Issues:0Issues:0

ipc

[Start here!] Flow-IPC - Modern C++ toolkit for high-speed inter-process communication (IPC)

Language:C++License:Apache-2.0Stargazers:233Issues:0Issues:0

llm.c

LLM training in simple, raw C/CUDA

Language:CudaLicense:MITStargazers:18833Issues:0Issues:0

ns3-ai

Enable the interaction between ns-3 and popular frameworks using Python, which mean you can train and test your AI algorithms in ns-3 without changing any frameworks you are using now!

Language:C++License:GPL-2.0Stargazers:202Issues:0Issues:0

Chinese-Resume-in-Typst

使用 Typst 编写的中文简历, 语法简洁, 样式美观, 开箱即用, 可选是否显示照片

Language:TypstStargazers:187Issues:0Issues:0

fms-fsdp

Demonstrate throughput of PyTorch FSDP

Language:PythonLicense:Apache-2.0Stargazers:70Issues:0Issues:0

makeMoE

From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)

Language:Jupyter NotebookLicense:MITStargazers:517Issues:0Issues:0
Language:C++License:MITStargazers:121Issues:0Issues:0

scattermoe

Triton-based implementation of Sparse Mixture of Experts.

Language:PythonLicense:Apache-2.0Stargazers:135Issues:0Issues:0

how-to-optim-algorithm-in-cuda

how to optimize some algorithm in cuda.

Language:CudaStargazers:1023Issues:0Issues:0

LapisCV

📃 开箱即用的 Obsidian / Typora 简历

Language:CSSLicense:MITStargazers:1500Issues:0Issues:0

text-clustering

Easily embed, cluster and semantically label text datasets

Language:PythonLicense:Apache-2.0Stargazers:345Issues:0Issues:0

LlamaGym

Fine-tune LLM agents with online reinforcement learning

Language:PythonLicense:MITStargazers:916Issues:0Issues:0

LLMs-from-scratch

Implementing a ChatGPT-like LLM from scratch, step by step

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:16401Issues:0Issues:0

hatchet

A distributed, fault-tolerant task queue

Language:GoLicense:MITStargazers:3294Issues:0Issues:0

tutel

Tutel MoE: An Optimized Mixture-of-Experts Implementation

Language:PythonLicense:MITStargazers:664Issues:0Issues:0

cookbook

Deep learning for dummies. All the practical details and useful utilities that go into working with real models.

Language:PythonLicense:Apache-2.0Stargazers:180Issues:0Issues:0

NewPipe

A libre lightweight streaming front-end for Android.

Language:JavaLicense:GPL-3.0Stargazers:28864Issues:0Issues:0

so-large-lm

大模型基础: 一文了解大模型基础知识

Stargazers:1159Issues:0Issues:0

glake

GLake: optimizing GPU memory management and IO transmission.

Language:C++License:Apache-2.0Stargazers:275Issues:0Issues:0

How_to_optimize_in_GPU

This is a series of GPU optimization topics. Here we will introduce how to optimize the CUDA kernel in detail. I will introduce several basic kernel optimizations, including: elementwise, reduce, sgemv, sgemm, etc. The performance of these kernels is basically at or near the theoretical limit.

Language:CudaLicense:Apache-2.0Stargazers:693Issues:0Issues:0

lectures

Material for cuda-mode lectures

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:835Issues:0Issues:0

nanotron

Minimalistic large language model 3D-parallelism training

Language:PythonLicense:Apache-2.0Stargazers:817Issues:0Issues:0