Nicolas Patry's repositories
alphagozero
Unofficial attempt to rebuild AlphaGo Zero
mistral.rs
Blazingly fast LLM inference.
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
Language:PythonApache-2.0000
Language:RustApache-2.0000
cudarc
Safe rust wrapper around CUDA toolkit
Language:RustApache-2.0000
doc-builder
The package used to build the documentation of our Hugging Face repos
Language:PythonApache-2.0000
exllamav2
A fast inference library for running LLMs locally on modern consumer-class GPUs
Language:PythonMIT000
flash-attention
Fast and memory-efficient exact attention
Language:PythonBSD-3-Clause000
flashinfer
FlashInfer: Kernel Library for LLM Serving
Language:CudaApache-2.0000
Language:Python000
Language:Nix000
llama.cpp
LLM inference in C/C++
Language:C++MIT000
narsil.github.io
My blog from fastpages
Language:HTMLCC-BY-SA-4.0000
tailscale-action
Github action to connect to tailscale
BSD-3-Clause000
Language:Nix000
Language:Rust000
000