There are 0 repository under cpu-inference topic.
Running Llama 2 and other Open-Source LLMs on CPU Inference Locally for Document Q&A
The bare metal in my basement
eLLM Infers LLM on CPUs in Real Time
Wrapper for simplified use of Llama2 GGUF quantized models.
Privacy-focused RAG chatbot for network documentation. Chat with your PDFs locally using Ollama, Chroma & LangChain. CPU-only, fully offline.
C# api wrapper for llm-inference chatllm.cpp
VB.NET api wrapper for llm-inference chatllm.cpp
🧠 A comprehensive toolkit for benchmarking, optimizing, and deploying local Large Language Models. Includes performance testing tools, optimized configurations for CPU/GPU/hybrid setups, and detailed guides to maximize LLM performance on your hardware.
Nim api-wrapper for llm-inference chatllm.cpp
PlantAi is a ResNet-based CNN model trained on the PlantVillage dataset to classify plant leaf images as healthy or diseased. This repository includes PyTorch training code, tools to convert the model to TensorFlow Lite (TFLite) for deployment, and an Android app integrating the model for real-time leaf disease detection from camera images.
Simple bot that transcribes Telegram voice messages. Powered by go-telegram-bot-api & whisper.cpp Go bindings.
rust api wrapper for llm-inference chatllm.cpp
kotlin api wrapper for llm-inference chatllm.cpp
lua api wrapper for llm-inference chatllm.cpp
gemma-2-2b-it int8 cpu inference in one file of pure C#
llama 3.2 1b fp16 cpu inference in one file of pure VB.NET
Java-port of qwen3.c
The Ark Project: Selecting the perfect AI model to reboot civilization from a 64GB USB drive. Comprehensive analysis of open-source LLMs under extreme constraints, with final recommendation: Meta Llama 3.1 70B Instruct (Q6_K GGUF). Includes interactive tools, detailed comparisons, and complete implementation guide for offline deployment.
🤖 AI Text Completion App built with Streamlit and Llama-3.2-1B. Generate creative text completions with an intuitive web interface. GPU & CPU optimized, easy to deploy, perfect for content creation and AI experimentation.
Un sistema RAG per chattare con documenti locali usando Foundry e modelli LLM su CPU
Lightweight web UI for llama.cpp with dynamic model switching, chat history & markdown support. No GPU required. Perfect for local AI development.