open-webui / open-webui

User-friendly WebUI for LLMs (Formerly Ollama WebUI)

Home Page:https://openwebui.com

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

enhancement: external stt

Stargate256 opened this issue · comments

Is your feature request related to a problem? Please describe.
Would it be possible to allow running whisper on another server?

Describe the solution you'd like
Adding the ability to connect to whisper via some kind of API.

Additional context
I am running open-webui on a server without GPUs, and LLMs, TTS on another server.

I've recently developed faster-whisper-server which provides an OpenAI compatible transcription/translation (and some extras as well) which could be used as a transcription backend if OpenWebUI adds support for having an external STT

@colaber2

STT & TTS settings have been refactored on dev and you can now configure external OpenAI STT from Admin Settings!

image

image

I tried this and the model name for STT is not correctly used.
Screenshot 2024-06-08 at 16-30-36 Admin Panel Open WebUI

ValueError: Invalid model size 'whisper-1', expected one of: tiny.en, tiny, base.en, base, small.en, small, medium.en, medium, large-v1, large-v2, large-v3, large, distil-large-v2, distil-medium.en, distil-small.en, distil-large-v3

@colaber2

STT & TTS settings have been refactored on dev and you can now configure external OpenAI STT from Admin Settings!

image

image

Amazing.. well done!!

I tried this and the model name for STT is not correctly used. Screenshot 2024-06-08 at 16-30-36 Admin Panel Open WebUI

ValueError: Invalid model size 'whisper-1', expected one of: tiny.en, tiny, base.en, base, small.en, small, medium.en, medium, large-v1, large-v2, large-v3, large, distil-large-v2, distil-medium.en, distil-small.en, distil-large-v3

@tjbck The model name is being ignored. As a result we can't use @fedirz's faster-whisper-server because it doesn't know what "whisper-1" is.
I modified faster-whisper-server to treat "whisper-1" as the default model and it worked great! @fedirz, maybe this is a good feature you could add? @tjbck, could you make the fix so we can choose the models?

Thank you both for your work!

@JohnSmithToYou yeah, I'll add special handling for whisper-1