OctoAI's repositories
octoai-textgen-cookbook
Simple getting-started code examples for LLM applications powered by OctoAI
octoml-llm-qa
A code sample that shows how to use 🦜️🔗langchain, 🦙llama_index and a hosted LLM endpoint to do a standard chat or Q&A about a pdf document
dockercon23-octoai
DockerCon 2023 OctoAI AI/ML Workshop GitHub Repo
LLM-RAG-Examples
OctoAI LLM RAG samples
octoai-apps
A collection of OctoAI-based demos.
macho-dyld
Custom dyld version inherited from original Apple dyld implementation
fern-config
Configuration for generating SDKs and Documentation.
llama-recipes
Examples and recipes for Llama 2 model
BiRefNet
[arXiv'24] Bilateral Reference for High-Resolution Dichotomous Image Segmentation
docker_auth
Authentication server for Docker Registry 2
homebrew-tap
Homebrew Tap of OctoML products and tools.
image-background-remove-tool
✂️ Automated high-quality background removal framework for an image using neural networks. ✂️
llm-examples
A repo containing code examples that feature OctoAI's LLM solution
octoai-model-examples
A set of models you can build and deploy on octoai
pinecone-rag-demo
Pinecone + Vercel RAG application, showcasing a comparison between chat with no context and using a Pinecone index for context
pre-commit-kustomize
pre-commit hook which runs kustomize docker image (use with https://github.com/pre-commit/pre-commit)
TensorRT-LLM-release
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
tflint-ruleset-google
TFLint ruleset for terraform-provider-google
unstructured
Open source libraries and APIs to build custom preprocessing pipelines for labeling, training, or production machine learning pipelines.
vllm-project
A high-throughput and memory-efficient inference and serving engine for LLMs