Mistral AI's repositories
mistral-src
Reference implementation of Mistral AI 7B v0.1 model.
client-python
Python client library for Mistral AI platform
vllm-release
A high-throughput and memory-efficient inference and serving engine for LLMs
FastChat-release
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
transformers-release
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.