KPREDDIE's repositories
gorillaAPIStoreFWD
Gorilla: An API store for LLMs
gpt-pilotAIDev2024
The first real AI developer
OS-Copilot
An self-improving embodied conversational agent seamlessly integrated into the operating system to automate our daily tasks.
tensorflow032024
An Open Source Machine Learning Framework for Everyone
auto-code-rover042024
A project structure aware autonomous software engineer aiming for autonomous program improvement. Resolved 15.95% tasks in full SWE-bench
auto-devAI
🧙AutoDev: The AI-powered coding wizard with multilingual support 🌐, auto code generation 🏗️, and a helpful bug-slaying assistant 🐞! Customizable prompts 🎨 and a magic Auto Dev/Testing/Document/Agent feature 🧪 included! 🚀
azure-search-openai-JSCRIPT2024
A TypeScript sample app for the Retrieval Augmented Generation pattern running on Azure, using Azure AI Search for retrieval and Azure OpenAI and LangChain large language models (LLMs) to power ChatGPT-style and Q&A experiences.
genai-stack022224
Langchain + Docker + Neo4j + Ollama
LLaVA-Interactive-DemoMultiM
LLaVA-Interactive-Demo
llm-answer-engine-FStack-RAG
Build a Perplexity-Inspired Answer Engine Using Next.js, Groq, Mixtral, Langchain, OpenAI, Brave & Serper
MemGPT042024
Building persistent LLM agents with long-term memory 📚🦙
MoraOpenSoraVariant
Mora: More like Sora for Generalist Video Generation
ollama022224
Get up and running with Llama 2, Mistral, and other large language models.
OpenDevinAPR2024
🐚 OpenDevin: Code Less, Make More
sf-samples-tastytest
Sample files, code snippets and downloads for Snowflake labs and tutorials.
singlefullstack-starter
NextJS, Prisma, SingleStore starter
singlestore-wasm-toolkitAPR2024
Tools to streamline development of Wasm UDFs for SingleStoreDB.
SWE-agent2024
SWE-agent takes a GitHub issue and tries to automatically fix it, using GPT-4, or your LM of choice. It solves 12.29% of bugs in the SWE-bench evaluation set and takes just 1.5 minutes to run.
TensorRT-LLMwinX
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
trt-llm-rag-windowsNVDA
A developer reference project for creating Retrieval Augmented Generation (RAG) chatbots on Windows using TensorRT-LLM