OctoAI's repositories
octoai-textgen-cookbook
Simple getting-started code examples for LLM applications powered by OctoAI
octoml-llm-qa
A code sample that shows how to use 🦜️🔗langchain, 🦙llama_index and a hosted LLM endpoint to do a standard chat or Q&A about a pdf document
dockercon23-octoai
DockerCon 2023 OctoAI AI/ML Workshop GitHub Repo
LLM-RAG-Examples
OctoAI LLM RAG samples
octoai-apps
A collection of OctoAI-based demos.
fern-config
Configuration for generating SDKs.
axolotl
Go ahead and axolotl questions
docker_auth
Authentication server for Docker Registry 2
go-jose
An implementation of JOSE standards (JWE, JWS, JWT) in Go
go-oidc
A Go OpenID Connect client.
homebrew-tap
Homebrew Tap of OctoML products and tools.
image-background-remove-tool
✂️ Automated high-quality background removal framework for an image using neural networks. ✂️
llama-recipes
Examples and recipes for Llama 2 model
llm-examples
A repo containing code examples that feature OctoAI's LLM solution
octoai-model-examples
A set of models you can build and deploy on octoai
pinecone-rag-demo
Pinecone + Vercel RAG application, showcasing a comparison between chat with no context and using a Pinecone index for context
pre-commit-kustomize
pre-commit hook which runs kustomize docker image (use with https://github.com/pre-commit/pre-commit)
TensorRT-LLM-release
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
tflint-ruleset-google
TFLint ruleset for terraform-provider-google
unstructured
Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.
use-whisper
React hook for OpenAI Whisper with speech recorder, real-time transcription, and silence removal built-in
vllm-project
A high-throughput and memory-efficient inference and serving engine for LLMs