There are 0 repository under model-inference-service topic.
The easiest way to serve AI apps and models - Build Model Inference APIs, Job queues, LLM apps, Multi-model pipelines, and more!
CLIP as a service - Embed image and sentences, object recognition, visual reasoning, image classification and reverse image search
Online Inference API for NLP Transformer models - summarization, text classification, sentiment analysis and more
Learn the ins and outs of efficiently serving Large Language Models (LLMs). Dive into optimization techniques, including KV caching and Low Rank Adapters (LoRA), and gain hands-on experience with Predibase’s LoRAX framework inference server.
SPIRA Serving Predictor v1 by @daitamae and @vitorguidi