Min Jun Kim's starred repositories

transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Language:PythonLicense:Apache-2.0Stargazers:134319Issues:1125Issues:16054

audiocraft

Audiocraft is a library for audio processing and generation with deep learning. It features the state-of-the-art EnCodec audio compressor / tokenizer, along with MusicGen, a simple and controllable music generation LM with textual and melodic conditioning.

Language:PythonLicense:MITStargazers:20877Issues:203Issues:381

unsloth

Finetune Llama 3.2, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory

Language:PythonLicense:Apache-2.0Stargazers:17682Issues:122Issues:957

AnimateAnyone

Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation

flash-attention

Fast and memory-efficient exact attention

Language:PythonLicense:BSD-3-ClauseStargazers:14005Issues:118Issues:1095

text-generation-inference

Large Language Model Text Generation Inference

Language:PythonLicense:Apache-2.0Stargazers:8987Issues:101Issues:1343

pyannote-audio

Neural building blocks for speaker diarization: speech activity detection, speaker change detection, overlapped speech detection, speaker embedding

Language:Jupyter NotebookLicense:MITStargazers:6239Issues:71Issues:992

mlx-examples

Examples in the MLX framework

Language:PythonLicense:MITStargazers:6128Issues:71Issues:495

MobileSAM

This is the official code for MobileSAM project that makes SAM lightweight for mobile applications and beyond!

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:4787Issues:43Issues:125

alignment-handbook

Robust recipes to align language models with human and AI preferences

Language:PythonLicense:Apache-2.0Stargazers:4636Issues:111Issues:135

CTranslate2

Fast inference engine for Transformer models

Language:C++License:MITStargazers:3370Issues:59Issues:702

lectures

Material for cuda-mode lectures

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:2487Issues:35Issues:7

priompt

Prompt design using JSX.

Language:TypeScriptLicense:MITStargazers:2000Issues:23Issues:7

RealtimeSTT

A robust, efficient, low-latency speech-to-text library with advanced voice activity detection, wake word activation and instant transcription.

Language:PythonLicense:MITStargazers:1973Issues:31Issues:94

RealtimeTTS

Converts text to speech in realtime

whisper-plus

WhisperPlus: Faster, Smarter, and More Capable 🚀

Language:PythonLicense:Apache-2.0Stargazers:1712Issues:19Issues:51

FlexFlow

FlexFlow Serve: Low-Latency, High-Performance LLM Serving

Language:C++License:Apache-2.0Stargazers:1695Issues:32Issues:659

coffee

Build and iterate on your UI 10x faster with AI - right from your own IDE ☕️

Language:PythonLicense:Apache-2.0Stargazers:1467Issues:8Issues:7

self-rewarding-lm-pytorch

Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI

Language:PythonLicense:MITStargazers:1328Issues:23Issues:17
Language:PythonLicense:Apache-2.0Stargazers:1252Issues:16Issues:115

AQLM

Official Pytorch repository for Extreme Compression of Large Language Models via Additive Quantization https://arxiv.org/pdf/2401.06118.pdf and PV-Tuning: Beyond Straight-Through Estimation for Extreme LLM Compression https://arxiv.org/abs/2405.14852

Language:PythonLicense:Apache-2.0Stargazers:1162Issues:19Issues:85

aphrodite-engine

Large-scale LLM inference engine

Language:PythonLicense:AGPL-3.0Stargazers:1091Issues:15Issues:183

react-native-skottie

▶️ Efficient lottie animations using Skia's Skottie module

Language:C++License:MITStargazers:867Issues:9Issues:40

VTubeStudio

VTube Studio API Development Page

Language:C#License:MITStargazers:849Issues:43Issues:57

talk

Let's make sand talk

tabbyAPI

An OAI compatible exllamav2 API that's both lightweight and fast

Language:PythonLicense:AGPL-3.0Stargazers:563Issues:10Issues:113

landmark-attention

Landmark Attention: Random-Access Infinite Context Length for Transformers

Language:PythonLicense:Apache-2.0Stargazers:414Issues:40Issues:15

lightspeedGPT

Use GPT4 and GPT3.5 on inputs of unlimited size. Uses multithreading to process multiple chunks in parallel. Useful for tasks like Named Entity Recognition, information extraction on large books, datasets, etc.

Language:Jupyter NotebookLicense:MITStargazers:272Issues:7Issues:1

landmark-attention-qlora

Landmark Attention: Random-Access Infinite Context Length for Transformers QLoRA

Language:PythonLicense:Apache-2.0Stargazers:124Issues:5Issues:0
Language:PythonLicense:GPL-3.0Stargazers:14Issues:0Issues:0