There are 20 repositories under llamacpp topic.
The all-in-one Desktop & Docker AI application with full RAG and AI Agent capabilities.
Private & local AI personal knowledge management app.
A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API.
Your AI second brain. A copilot to get answers to your questions, whether they be from your own notes or from the internet. Use powerful, online (e.g gpt4) or private, local (e.g mistral) LLMs. Self-host locally or use our web app. Access from Obsidian, Emacs, Desktop app, Web or Whatsapp.
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
AGiXT is a dynamic AI Agent Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.
A C#/.NET library to run LLM models (🦙LLaMA/LLaVA) on your local device efficiently.
The most no-nonsense, locally or API-hosted AI code completion plugin for Visual Studio Code - like GitHub Copilot but completely free and 100% private.
Chat with your favourite LLaMA models in a native macOS app
The TypeScript library for building AI applications.
A FastAPI service for semantic text search using precomputed embeddings and advanced similarity measures, with built-in support for various file types through textract.
Believe in AI democratization. llama for nodejs backed by llama-rs, llama.cpp and rwkv.cpp, work locally on your laptop CPU. support llama/alpaca/gpt4all/vicuna/rwkv model.
Maid is a cross-platform Flutter app for interfacing with GGUF / llama.cpp models locally, and with Ollama and OpenAI models remotely.
Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization
Use Code Llama with Visual Studio Code and the Continue extension. A local LLM alternative to GitHub Copilot.
A simple "Be My Eyes" web app with a llama.cpp/llava backend
llama.cpp with BakLLaVA model describes what does it see
Local first semantic code search and chat powered by vector embeddings and LLMs
The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM models, execute structured function calls and get structured output. Works also with models not fine-tuned to JSON output and function calls.
The "vicuna-installation-guide" provides step-by-step instructions for installing and configuring Vicuna 13 and 7B
An innovative library for efficient LLM inference via low-bit quantization
AI-powered cybersecurity chatbot designed to provide helpful and accurate answers to your cybersecurity-related queries and also do code analysis and scan analysis.
A suite of custom nodes for ConfyUI that includes GPT text-prompt generation, LoadVideo, SaveVideo, LoadFramesFromFolder and FrameInterpolator
AubAI brings you on-device gen-AI capabilities, including offline text generation and more, directly within your app.