OctoAI (octoml)

OctoAI

octoml

Geek Repo

Optimizing machine learning using machine learning

Location:Seattle

Home Page:octo.ai

Github PK Tool:Github PK Tool

OctoAI's repositories

octoml-profile

Home for OctoML PyTorch Profiler

triton-client-rs

A client library in Rust for Nvidia Triton.

Language:RustLicense:Apache-2.0Stargazers:23Issues:37Issues:2

octoml-llm-qa

A code sample that shows how to use 🦜️🔗langchain, 🦙llama_index and a hosted LLM endpoint to do a standard chat or Q&A about a pdf document

relax

A fork of tvm/unity

Language:PythonLicense:Apache-2.0Stargazers:16Issues:7Issues:0

tvm2onnx

An open-source tool created by OctoML that converts TVM-optimized models to code runnable in ONNX Runtime.

Language:PythonLicense:Apache-2.0Stargazers:14Issues:32Issues:1

dockercon23-octoai

DockerCon 2023 OctoAI AI/ML Workshop GitHub Repo

Language:Jupyter NotebookStargazers:8Issues:3Issues:0
Language:CLicense:Apache-2.0Stargazers:8Issues:40Issues:1

mlc-llm

Enable everyone to develop, optimize and deploy AI models natively on everyone's devices.

Language:PythonLicense:Apache-2.0Stargazers:5Issues:1Issues:16

octoai-apps

A collection of OctoAI-based demos.

Language:TypeScriptLicense:Apache-2.0Stargazers:5Issues:28Issues:1

hackathon-2023-rag

OctoAI 2023 Llama2 RAG demos

Language:PythonStargazers:4Issues:0Issues:0
Language:PythonLicense:NOASSERTIONStargazers:3Issues:3Issues:0

inference

Reference implementations of MLPerf™ inference benchmarks

Language:PythonLicense:NOASSERTIONStargazers:2Issues:1Issues:0

octoai-cartoonizer

Cartoonizer demo for OctoAI compute service launch

Language:PythonStargazers:1Issues:1Issues:0

octoai-launch-examples

Examples of how to build Generative AI applications powered by the OctoAI compute service.

Language:Jupyter NotebookStargazers:1Issues:4Issues:0

langchain

⚡ Building applications with LLMs through composability ⚡

Language:PythonLicense:MITStargazers:0Issues:1Issues:0

archived_vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:3Issues:0

homebrew-tap

Homebrew Tap of OctoML products and tools.

Language:RubyLicense:Apache-2.0Stargazers:0Issues:4Issues:0
Language:PythonStargazers:0Issues:3Issues:0

octoai-octoshop

OctoAI's OctoShop! Transform photos with the power of words and generative AI!

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

power-dev

Dev repo for power measurement for the MLPerf™ benchmarks

Language:PythonLicense:NOASSERTIONStargazers:0Issues:1Issues:0

relax-all

A fork of tvm/unity

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:3Issues:0

stable-diffusion-webui-docker

Easy Docker setup for Stable Diffusion with user-friendly UI

Language:DockerfileLicense:NOASSERTIONStargazers:0Issues:1Issues:0

TensorRT-LLM-release

TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.

Language:C++License:Apache-2.0Stargazers:0Issues:1Issues:0

triton-inference-server

The Triton Inference Server provides an optimized cloud and edge inferencing solution.

Language:PythonLicense:BSD-3-ClauseStargazers:0Issues:0Issues:0

use-whisper

React hook for OpenAI Whisper with speech recorder, real-time transcription, and silence removal built-in

Language:TypeScriptLicense:MITStargazers:0Issues:1Issues:0

web-llm

Bringing large-language models and chat to web browsers. Everything runs inside the browser with no server support.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0