jllllll's repositories
bitsandbytes-windows-webui
Windows compile of bitsandbytes for use in text-generation-webui.
llama-cpp-python-cuBLAS-wheels
Wheels for llama-cpp-python compiled with cuBLAS support
one-click-installers
Simplified installers for oobabooga/text-generation-webui.
bitsandbytes
8-bit CUDA functions for PyTorch
GPTQ-for-LLaMa-Wheels
Precompiled Wheels for GPTQ-for-LLaMa
flash-attention
Fast and memory-efficient exact attention - Windows wheels
GPTQ-for-LLaMa-CUDA
A combination of Oobabooga's fork and the main cuda branch of GPTQ-for-LLaMa in a package format.
ctransformers-cuBLAS-wheels
ctransformers wheels with pre-built CUDA binaries for additional CUDA and AVX versions.
windows-venv-installers
Standalone, dependency-less scripts for automatically setting up a virtual environment for easy project installation on Windows.
text-generation-webui
A gradio web UI for running Large Language Models like LLaMA, llama.cpp, GPT-J, Pythia, OPT, and GALACTICA.
GPTQ-for-LLaMa
4 bits quantization of LLMs using GPTQ
ctransformers
Python bindings for the Transformer models implemented in C/C++ using GGML library.
safetensors
Simple, safe way to store and distribute tensors
scikit-build-core
A next generation Python CMake adaptor and Python API for plugins
SillyTavern
LLM Frontend for Power Users.