Baseten's repositories
truss-examples
Examples of models deployable with Truss
starcoder-truss
Truss for deploying Starcoder to Baseten or other platforms
falcon-7b-truss
Falcon 7b
infrastructure-take-home
Baseten infrastructure recruiting take home
ControlNet
Let us control diffusion models
pygmalion-6b-truss
A Truss to deploy Pygmalion 6B on Baseten.
chainlit-cookbook
Chainlit's cookbook repo
diffusers
🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch
gpu-operator
NVIDIA GPU Operator creates/configures/manages GPUs atop Kubernetes
kaniko
Build Container Images In Kubernetes
langchain
⚡ Building applications with LLMs through composability ⚡
mpt-7b-base-truss
A deployment "truss" for the MPT-7B Base model from MosaicML
python_backend
Triton backend that enables pre-process, post-processing and other logic to be implemented in Python.
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
tensorrtllm_backend
The Triton TensorRT-LLM Backend
triton-inference-server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
truss-public-gh-repo-test
A public github repo for testing truss deploy flow