oboje / whisper-ui

Streamlit UI for OpenAI's Whisper

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Streamlit UI for OpenAI's Whisper

This is a simple Streamlit UI for OpenAI's Whisper speech-to-text model. It let's you download and transcribe media from YouTube videos, playlists, or local files. You can then browse, filter, and search through your saved audio files. Feel free to raise an issue for bugs or feature requests or send a PR.

whisper-ui-update-demo.mp4

Setup

This was built & tested on Python 3.11 but should also work on Python 3.9+ as with the original Whisper repo). You'll need to install ffmpeg on your system. Then, install the requirements with pip.

sudo apt install ffmpeg
pip install -r requirements.txt

If you're using conda, you can create a new environment with the following command:

conda env create -f environment.yml

Note: If you're using a CPU-only machine, your runtime can be sped-up by using quantization implemented by @MicellaneousStuff by swapping out pip install openai-whisper from requirements.txt and replacing it with their fork pip install git+https://github.com/MiscellaneousStuff/whisper.git (See related discussion here - hayabhay#20)

Usage

Once you're set up, you can run the app with:

streamlit run app/01_🏠_Home.py

This will open a new tab in your browser with the app. You can then select a YouTube URL or local file & click "Run Whisper" to run the model on the selected media.

Changelog

All notable changes to this project alongside potential feature roadmap will be documented in this file.

License

Whisper is licensed under MIT while Streamlit is licensed under Apache 2.0. Everything else is licensed under MIT.

About

Streamlit UI for OpenAI's Whisper

License:MIT License


Languages

Language:Python 100.0%