EmbeddedLLM / vllm-rocm

vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs

Home Page:https://vllm.readthedocs.io

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

EmbeddedLLM/vllm-rocm Stargazers