There are 31 repositories under llamacpp topic.
Jan is an open source alternative to ChatGPT that runs 100% offline on your computer
Your AI second brain. Self-hostable. Get answers from the web or your docs. Build custom agents, schedule automations, do deep research. Turn any online or local LLM into your personal, autonomous AI (gpt, claude, gemini, llama, qwen, mistral). Get started - free.
Unified framework for building enterprise RAG pipelines with small, specialized models
A self-hosted, offline, ChatGPT-like chatbot. Powered by Llama 2. 100% private, with no data leaving your device. New: Code Llama support!
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
Private & local AI personal knowledge management app for high entropy people.
A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API.
State of the Art Natural Language Processing
Your agent in your terminal, equipped with local tools: writes code, uses the terminal, browses the web, vision.
The most no-nonsense, locally or API-hosted AI code completion plugin for Visual Studio Code - like GitHub Copilot but 100% free.
A C#/.NET library to run LLM (🦙LLaMA/LLaVA) on your local device efficiently.
LSP-AI is an open-source language server that serves as a backend for AI-powered functionality, designed to assist and empower software engineers, not replace them.
AGiXT is a dynamic AI Agent Automation Platform that seamlessly orchestrates instruction management and complex task execution across diverse AI providers. Combining adaptive memory, smart features, and a versatile plugin system, AGiXT delivers efficient and comprehensive AI solutions.
Maid is a cross-platform Flutter app for interfacing with GGUF / llama.cpp models locally, and with Ollama and OpenAI models remotely.
The goal of RamaLama is to make working with AI boring.
Chat with your favourite LLaMA models in a native macOS app
Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization
The TypeScript library for building AI applications.
The easiest way to use Ollama in .NET
A FastAPI service for semantic text search using precomputed embeddings and advanced similarity measures, with built-in support for various file types through textract.
Believe in AI democratization. llama for nodejs backed by llama-rs, llama.cpp and rwkv.cpp, work locally on your laptop CPU. support llama/alpaca/gpt4all/vicuna/rwkv model.
LSP server leveraging LLMs for code completion (and more?)
Practical Llama 3 inference in Java
Stateful load balancer custom-tailored for llama.cpp 🏓🦙
ComfyUI-IF_AI_tools is a set of custom nodes for ComfyUI that allows you to generate prompts using a local Large Language Model (LLM) via Ollama. This tool enables you to enhance your image generation workflow by leveraging the power of language models.
Use Code Llama with Visual Studio Code and the Continue extension. A local LLM alternative to GitHub Copilot.
The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM models, execute structured function calls and get structured output. Works also with models not fine-tuned to JSON output and function calls.
Model swapping for llama.cpp (or any local OpenAPI compatible server)
A simple "Be My Eyes" web app with a llama.cpp/llava backend
Local first semantic code search and chat | Leverage custom copilots with fine-tuning datasets from code in Alpaca, Conversational, Completion and Instruction format