ChaosCodes's starred repositories

mini-omni

open-source multimodal large language model that can hear, talk while thinking. Featuring real-time end-to-end speech input and streaming audio output conversational capabilities.

Language:PythonLicense:MITStargazers:2745Issues:0Issues:0

flute

Fast Matrix Multiplications for Lookup Table-Quantized LLMs

Language:CudaLicense:Apache-2.0Stargazers:168Issues:0Issues:0

attention-gym

Helpful tools and examples for working with flex-attention

Language:PythonLicense:BSD-3-ClauseStargazers:371Issues:0Issues:0

MiniCPM-V

MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone

Language:PythonLicense:Apache-2.0Stargazers:12139Issues:0Issues:0

torchtitan

A native PyTorch Library for large model training

Language:PythonLicense:BSD-3-ClauseStargazers:2310Issues:0Issues:0
Stargazers:25Issues:0Issues:0

Knowledge-Conflicts-Survey

[EMNLP 2024] The official GitHub repo for the survey paper "Knowledge Conflicts for LLMs: A Survey"

Stargazers:66Issues:0Issues:0

TensorRT-Model-Optimizer

TensorRT Model Optimizer is a unified library of state-of-the-art model optimization techniques such as quantization, pruning, distillation, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM or TensorRT to optimize inference speed on NVIDIA GPUs.

Language:PythonLicense:NOASSERTIONStargazers:459Issues:0Issues:0

ThunderKittens

Tile primitives for speedy kernels

Language:CudaLicense:MITStargazers:1516Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:107Issues:0Issues:0

qserve

QServe: W4A8KV4 Quantization and System Co-design for Efficient LLM Serving

Language:PythonLicense:Apache-2.0Stargazers:407Issues:0Issues:0

flash-linear-attention

Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton

Language:PythonLicense:MITStargazers:1243Issues:0Issues:0

Steel-LLM

Train a Chinese LLM From 0 by Personal

Language:PythonStargazers:153Issues:0Issues:0

llm.c

LLM training in simple, raw C/CUDA

Language:CudaLicense:MITStargazers:23712Issues:0Issues:0

EasyContext

Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.

Language:PythonLicense:Apache-2.0Stargazers:621Issues:0Issues:0

x-transformers

A concise but complete full-attention transformer with a set of promising experimental features from various papers

Language:PythonLicense:MITStargazers:4623Issues:0Issues:0

scattermoe

Triton-based implementation of Sparse Mixture of Experts.

Language:PythonLicense:Apache-2.0Stargazers:170Issues:0Issues:0

lmms-eval

Accelerating the development of large multimodal models (LMMs) with lmms-eval

Language:PythonLicense:NOASSERTIONStargazers:1405Issues:0Issues:0

LLaMA-Factory

Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)

Language:PythonLicense:Apache-2.0Stargazers:31870Issues:0Issues:0

sailor-llm

⚓️ Sailor: Open Language Models for South-East Asia

Language:PythonLicense:MITStargazers:104Issues:0Issues:0

MergeLM

Codebase for Merging Language Models (ICML 2024)

Language:PythonStargazers:749Issues:0Issues:0

DataDreamer

DataDreamer: Prompt. Generate Synthetic Data. Train & Align Models.   🤖💤

Language:PythonLicense:MITStargazers:809Issues:0Issues:0

text-dedup

All-in-one text de-duplication

Language:PythonLicense:Apache-2.0Stargazers:593Issues:0Issues:0

sglang

SGLang is a fast serving framework for large language models and vision language models.

Language:PythonLicense:Apache-2.0Stargazers:5409Issues:0Issues:0

datatrove

Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.

Language:PythonLicense:Apache-2.0Stargazers:1971Issues:0Issues:0

gradio

Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work!

Language:PythonLicense:Apache-2.0Stargazers:32515Issues:0Issues:0

llama-moe

⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training (EMNLP 2024)

Language:PythonLicense:Apache-2.0Stargazers:861Issues:0Issues:0

tangent_task_arithmetic

Source code of "Task arithmetic in the tangent space: Improved editing of pre-trained models".

Language:PythonLicense:MITStargazers:81Issues:0Issues:0
Language:PythonLicense:MITStargazers:2491Issues:0Issues:0

trl

Train transformer language models with reinforcement learning.

Language:PythonLicense:Apache-2.0Stargazers:9604Issues:0Issues:0