mistralai / vllm-release

A high-throughput and memory-efficient inference and serving engine for LLMs

Home Page:https://vllm.readthedocs.io

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

mistralai/vllm-release Stargazers