Georgi Gerganov's starred repositories
LLocalSearch
LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a chain of LLMs to find the answer. The user can see the progress of the agents and the final answer. No OpenAI or Google API keys are needed.
distributed-llama
Tensor parallelism is all you need. Run LLMs on weak devices or make powerful devices even more powerful by distributing the workload and dividing the RAM usage.
openvino-plugins-ai-audacity
A set of AI-enabled effects, generators, and analyzers for Audacity®.
llama-cpp-agent
The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM models, execute structured function calls and get structured output. Works also with models not fine-tuned to JSON output and function calls.
chatllm.cpp
Pure C++ implementation of several models for real-time chatting on your computer (CPU)
llama-cpp-wasm
WebAssembly (Wasm) Build and Bindings for llama.cpp
llama_cpp-rs
High-level, optionally asynchronous Rust bindings to llama.cpp
rk3588-npu
Reverse engineering the rk3588 npu
whispercpp
Pybind11 bindings for Whisper.cpp
emoji_finder
emoji_finder