Python tools for WhisperKit
- Convert PyTorch Whisper models to WhisperKit format
- Apply custom inference optimizations and model compression
- Evaluate Whisper using WhisperKit and other Whisper implementations on benchmarks
- Installation
- Model Generation
- Model Evaluation
- Python Inference
- Example SwiftUI App
- Quality-of-Inference
- FAQ
- Citation
- Step 1: Fork this repository
- Step 2: Create a Python virtual environment, e.g.:
conda create -n whisperkit python=3.11 -y && conda activate whisperkit
- Step 3: Install as editable
cd WHISPERKIT_ROOT_DIR && pip install -e .
Convert Hugging Face Whisper Models (PyTorch) to WhisperKit (Core ML) format:
whisperkit-generate-model --model-version <model-version> --output-dir <output-dir>
For optional arguments related to model optimizations, please see the help menu with -h
We host several popular Whisper model versions here. These hosted models are automatically over-the-air deployable to apps integrating WhisperKit such as our example app WhisperAX on TestFlight. If you would like to publish custom Whisper versions that are not already published, you can do so as follows:
- Step 1: Find the user or organization name that you have write access to on Hugging Face Hub. If you are logged into
huggingface-cli
locally, you may simply do:
huggingface-cli whoami
If you don't have a write token yet, you can generate it here.
- Step 2: Point to the model repository that you would like to publish to, e.g.
my-org/my-whisper-repo-name
, with theMODEL_REPO_ID
environment variable and specify the name of the source PyTorch Whisper repository (e.g. distil-whisper/distil-small.en)
MODEL_REPO_ID=my-org/my-whisper-repo-name whisperkit-generate-model --model-version distil-whisper/distil-small.en --output-dir <output-dir>
If the above command is successfuly executed, your model will have been published to hf.co/my-org/my-whisper-repo-name/distil-whisper_distil-small.en
!
Evaluate (Argmax- or developer-published) models on speech recognition datasets:
whisperkit-evaluate-model --model-version <model-version> --output-dir <output-dir> --dataset {librispeech-debug,librispeech,earnings22}
By default, this command uses the latest main
branch commits from WhisperKit
and searches within Argmax-published model repositories. For optional arguments related to code and model versioning, please see the help menu with -h
We continually publish the evaluation results of Argmax-hosted models here as part of our continuous integration tests.
If you would like to evaluate WhisperKit models on your own dataset:
- Step 1: Publish a dataset on the Hub with the same simple structure as this toy dataset (audio files +
metadata.json
) - Step 2: Run evaluation with environment variables as follows:
export CUSTOM_EVAL_DATASET="my-dataset-name-on-hub"
export DATASET_REPO_OWNER="my-user-or-org-name-on-hub"
export MODEL_REPO_ID="my-org/my-whisper-repo-name" # if evaluating self-published models
whisperkit-evaluate-model --model-version <model-version> --output-dir <output-dir> --dataset my-dataset-name-on-hub
Use the unified Python wrapper for several Whisper frameworks:
from whisperkit.pipelines import WhisperKit, WhisperCpp, WhisperMLX
pipe = WhisperKit(whisper_version="openai/whisper-large-v3", out_dir="/path/to/out/dir")
print(pipe("audio.{wav,flac,mp3}"))
Note: Using WhisperCpp
requires ffmpeg
to be installed. Recommended installation is with brew install ffmpeg
This app serves two purposes:
- Base template for developers to freely customize and integrate parts into their own app
- Real-world testing/debugging utility for custom Whisper versions or WhisperKit features before/without building an app.
Note that the app is in beta and we are actively seeking feedback to improve it before widely distributing it.
We believe that rigorously measuring the quality of inference is necessary for developers and enterprises to make informed decisions when opting to use optimized or compressed variants of Whisper models in production. The current measurements are between reference and optimized WhisperKit models. We are going to extend the scope of this measurement to other Whisper implementations soon so developers can certify the behavior change (if any) caused by alternating use of WhisperKit with (or migration from) these implementations.
In all measurements, we care primarily about per-example no-regressions (quantified as qoi
below)
which is a stricter metric compared to dataset average WER. A 100% qoi
preserves perfect
backwards-compatibility on the test distribution and avoids "perceived regressions", the phenomenon
where per-example known behavior changes after a code/model update and causes divergence in
downstream code or breaks the user experience itself (even if dataset averages might stay flat
across updates). Pseudocode for qoi
:
qoi = []
for example in dataset:
no_regression = wer(optimized_model(example)) <= wer(reference_model(example))
qoi.append(no_regression)
qoi = (sum(qoi) / len(qoi)) * 100.
We define the reference model as the default float16 precision Core ML model that is generated by
whisperkittools. This reference model matches the accuracy of the original PyTorch model
on the specified test sets. We use librispeech/test.clean
(5 hours of short English audio clips)
as our testing set for Whisper. We are actively expanding our test set coverage to earnings22
(120 hours of long English audio clips with various accents). We anticipate developers that use Whisper in production to have
their own Quality Assurance test sets and whisperkittools offers the tooling necessary to run the
same measurements on such custom test sets, please see the Model Evaluation on Custom Dataset
for details.
Results in this page are generated by our cluster of Apple Silicon Macs. We use them as self-hosted runners on
Github Actions as our CI infrastructure. Due to security concerns,
we are unable to open up the cluster to the public. However, any Apple Silicon Mac (even with 8GB RAM) can be used to
run identical evaluation jobs
locally. For reference, our M2 Ultra devices complete a librispeech
+ openai/whisper-large-v3
evaluation in under 1 hour regardless of the Whisper implementation. Older Apple Silicon Macs should take less than
1 day to complete the same evaluation.
WER | QoI (%) | File Size (MB) | |
---|---|---|---|
openai_whisper-large-v3 | 2.44 | 100 | 3100 |
openai_whisper-large-v3_turbo | 2.54 | 99.5 | 3100 |
openai_whisper-large-v3_turbo_1128MB | 3.27 | 93 | 1128 |
openai_whisper-large-v3_turbo_1398MB | 2.57 | 97.8 | 1398 |
WER | Commit Hash | Model Format | |
---|---|---|---|
WhisperKit | 2.44 | 14e705e | Core ML |
WhisperCpp | 2.57 | 4bbb60e | Core ML + GGUF |
WhisperMLX | 2.57 | 854ad87 | MLX (Numpy) |
-
_turbo
: Indicates the presence of additional optimizations (not compression) to unlock streaming transcription as described in our Blog Post. -
_*MB
: Indicates the presence of mixed-bit quantization. Instead of cluttering the filename with details like_AudioEncoder-5.8bits_TextDecoder-6.1bits
, we choose to summarize the compression spec as the resulting total file size since this is what matters to developers in production.
Q1: xcrun: error: unable to find utility "coremlcompiler", not a developer tool or in PATH
A1: Ensure Xcode is installed on your Mac and run sudo xcode-select --switch /Applications/Xcode.app/Contents/Developer
.
If you use WhisperKit for something cool or just find it useful, please drop us a note at info@takeargmax.com!
If you use WhisperKit for academic work, here is the BibTeX:
@misc{whisperkit-argmax,
title = {WhisperKit},
author = {Argmax, Inc.},
year = {2024},
URL = {https://github.com/argmaxinc/WhisperKit}
}