enwask / ghostwheel-webui

Fork of Open WebUI for use with the ghostwheel inference server

Home Page:https://openwebui.com

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Open WebUI with ghostwheel

Overview

This repository provides a fork of Open WebUI with some hacky changes to the Ollama backend, to allow for use with the ghostwheel inference server. Most of the base functionality of WebUI should work out of the box, but certain admin operations (i.e. pulling, pushing, editing, deleting models from the Ollama server) are disabled when using ghostwheel. You'll receive an error message in the panel, and a 405 from the WebUI backend, when attempting to use these features.

Other functionality works as expected—you can chat with the existing models, create new ones from the base models with specific knowledge or prompts, embed documents and use them for RAG question/answer, etc.

Usage

You can set up and run WebUI from this fork, serving WebUI locally with ghostwheel as its Ollama backend. The instructions are very similar to (and derived from) WebUI's own guide for local deployment. Note that we've made some changes to the default environment configuration to disable user login (since you'll likely only be running locally), as well as to enable ghostwheel and provide it your API key.

Run the following commands to install WebUI with ghostwheel locally (you will of course need an appropriate Python environment, as well as an npm installation):

# Clone the repository locally and enter it
git clone https://github.com/enwask/ghostwheel-webui.git
cd ghostwheel-webui

# Copy the .env file
# You must set your ghostwheel API key in the .env file, once created
cp -a .env.example .env

# Build the frontend with Node
npm install
npm run build

# Install requirements for the WebUI backend
cd backend
pip install -Ur requirements.txt

# Finally, run the start script
# This script serves both the backend & frontend for us
bash start.sh

Setup for running in Docker, or in another production environment, is left as an exercise for the reader. It shouldn't differ too much from a standard setup for WebUI with a remote Ollama backend, besides using the relevant ghostwheel environment variables in .env (again, see .env.example).

About

Fork of Open WebUI for use with the ghostwheel inference server

https://openwebui.com

License:MIT License


Languages

Language:Svelte 59.1%Language:Python 27.5%Language:TypeScript 10.9%Language:Shell 0.8%Language:CSS 0.7%Language:Dockerfile 0.4%Language:JavaScript 0.3%Language:HTML 0.2%Language:Batchfile 0.1%Language:Makefile 0.0%