Peng Wu (pengwu22)

pengwu22

Geek Repo

Company:ByteDance

Location:Mountain View

Github PK Tool:Github PK Tool

Peng Wu's starred repositories

open-interpreter

A natural language interface for computers

Language:PythonLicense:AGPL-3.0Stargazers:52454Issues:0Issues:0

Mooncake

Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.

Stargazers:1066Issues:0Issues:0

Quest

[ICML 2024] Quest: Query-Aware Sparsity for Efficient Long-Context LLM Inference

Language:CudaStargazers:173Issues:0Issues:0

llm.c

LLM training in simple, raw C/CUDA

Language:CudaLicense:MITStargazers:23702Issues:0Issues:0

JetStream

JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs welcome).

Language:PythonLicense:Apache-2.0Stargazers:202Issues:0Issues:0

gemma_pytorch

The official PyTorch implementation of Google's Gemma models

Language:PythonLicense:Apache-2.0Stargazers:5248Issues:0Issues:0

DeepSeek-Coder

DeepSeek Coder: Let the Code Write Itself

Language:PythonLicense:MITStargazers:6620Issues:0Issues:0

basic-pitch

A lightweight yet powerful audio-to-MIDI converter with pitch bend detection

Language:PythonLicense:Apache-2.0Stargazers:3371Issues:0Issues:0

photoprism

AI-Powered Photos App for the Decentralized Web 🌈💎✨

Language:GoLicense:NOASSERTIONStargazers:34897Issues:0Issues:0

codellama

Inference code for CodeLlama models

Language:PythonLicense:NOASSERTIONStargazers:15925Issues:0Issues:0

BladeDISC

BladeDISC is an end-to-end DynamIc Shape Compiler project for machine learning workloads.

Language:C++License:Apache-2.0Stargazers:802Issues:0Issues:0

nccl

Optimized primitives for collective multi-GPU communication

Language:C++License:NOASSERTIONStargazers:3155Issues:0Issues:0

audiocraft

Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable music generation LM with textual and melodic conditioning.

Language:PythonLicense:MITStargazers:20696Issues:0Issues:0

bitsandbytes

Accessible large language models via k-bit quantization for PyTorch.

Language:PythonLicense:MITStargazers:6129Issues:0Issues:0

llama.cpp

LLM inference in C/C++

Language:C++License:MITStargazers:65780Issues:0Issues:0

vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

Language:PythonLicense:Apache-2.0Stargazers:27809Issues:0Issues:0

FastChat

An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.

Language:PythonLicense:Apache-2.0Stargazers:36584Issues:0Issues:0

stable-diffusion-webui

Stable Diffusion web UI

Language:PythonLicense:AGPL-3.0Stargazers:140644Issues:0Issues:0

ControlNet

Let us control diffusion models!

Language:PythonLicense:Apache-2.0Stargazers:29942Issues:0Issues:0

lora

Using Low-rank adaptation to quickly fine-tune diffusion models.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:6991Issues:0Issues:0

opennurbs

OpenNURBS libraries allow anyone to read and write the 3DM file format without the need for Rhino.

Language:C++License:NOASSERTIONStargazers:409Issues:0Issues:0

evals

Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.

Language:PythonLicense:NOASSERTIONStargazers:14760Issues:0Issues:0

FlexiGen

Running large language models on a single GPU for throughput-oriented scenarios.

Language:PythonLicense:Apache-2.0Stargazers:9149Issues:0Issues:0

byteir

A model compilation solution for various hardware

Language:MLIRLicense:Apache-2.0Stargazers:362Issues:0Issues:0

sheetsage

Transcribe music into lead sheets!

Language:PythonLicense:NOASSERTIONStargazers:301Issues:0Issues:0

triton

Development repository for the Triton language and compiler

Language:C++License:MITStargazers:12921Issues:0Issues:0

llama_index

LlamaIndex is a data framework for your LLM applications

Language:PythonLicense:MITStargazers:35857Issues:0Issues:0
Language:PythonStargazers:3464Issues:0Issues:0

DeepSpeed-MII

MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.

Language:PythonLicense:Apache-2.0Stargazers:1853Issues:0Issues:0

google-research

Google Research

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:33973Issues:0Issues:0