Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization
Home Page:https://rahulschand.github.io/gpu_poor/
Geek Repo:Geek Repo
Github PK Tool:Github PK Tool