anselbrandt / vllm-server

Repository from Github https://github.comanselbrandt/vllm-serverRepository from Github https://github.comanselbrandt/vllm-server

vLLM Server

Run

uv run vllm serve Qwen/Qwen2.5-3B-Instruct

Docker

docker build -t vllm-server .

docker run --privileged --gpus all -d -p 8000:8000 vllm-server

Test

curl http://localhost:8000/v1/completions \
    -H "Content-Type: application/json" \
    -d '{
        "model": "Qwen/Qwen2.5-3B-Instruct",
        "prompt": "What is the deal with peanut butter?",
        "max_tokens": 500,
        "temperature": 0
    }'

About


Languages

Language:Dockerfile 100.0%