Python tools for WhisperKit and WhisperKitAndroid
- Convert PyTorch Whisper models to WhisperKit format
- Apply custom inference optimizations and model compression
- Evaluate Whisper using WhisperKit and other Whisper implementations on benchmarks
Important
If you are looking for more features such as speaker diarization and upgraded performance, check out WhisperKit Pro!
- Installation
- Model Generation (Apple)
- Model Generation (Android)
- Model Evaluation (Apple)
- Python Inference
- Example SwiftUI App
- Quality-of-Inference
- FAQ
- Citation
- Step 1: Fork this repository
- Step 2: Create a Python virtual environment, e.g.:
conda create -n whisperkit python=3.11 -y && conda activate whisperkit
- Step 3: Install the base package as editable
cd WHISPERKIT_ROOT_DIR && pip install -e .
Convert Hugging Face Whisper Models (PyTorch) to WhisperKit (Core ML) format:
whisperkit-generate-model --model-version <model-version> --output-dir <output-dir>
For optional arguments related to model optimizations, please see the help menu with -h
We host several popular Whisper model versions here. These hosted models are automatically over-the-air deployable to apps integrating WhisperKit such as our example app WhisperAX on TestFlight. If you would like to publish custom Whisper versions that are not already published, you can do so as follows:
- Step 1: Find the user or organization name that you have write access to on Hugging Face Hub. If you are logged into
huggingface-cli
locally, you may simply do:
huggingface-cli whoami
If you don't have a write token yet, you can generate it here.
- Step 2: Point to the model repository that you would like to publish to, e.g.
my-org/my-whisper-repo-name
, with theMODEL_REPO_ID
environment variable and specify the name of the source PyTorch Whisper repository (e.g. distil-whisper/distil-small.en)
MODEL_REPO_ID=my-org/my-whisper-repo-name whisperkit-generate-model --model-version distil-whisper/distil-small.en --output-dir <output-dir>
If the above command is successfuly executed, your model will have been published to hf.co/my-org/my-whisper-repo-name/distil-whisper_distil-small.en
!
WhisperKit currently only supports Qualcomm AI Hub Whisper models on Hugging Face:
whisperkittools generates 3 more support models for input preprocessing and output postprocessing used in the WhisperKitAndroid pipeline. These are all published on Hugging Face here. Nonetheless, you may regenerate these models if you wish by following these steps:
- Step 1: Create an account at aihub.qualcomm.com
- Step 2: Set your API key locally as
qai-hub configure --api_token
- Step 3: Install extra dependencies via
pip install -e '.[android]'
(Note that this requirespython<3.11
) - Step 4: Execute
python tests/test_aihub.py --persistent-cache-dir <output-path>
Stay tuned for more options for generating models without creating an account and more model version coverage!
Evaluate (Argmax- or developer-published) models on speech recognition datasets:
whisperkit-evaluate-model --model-version <model-version> --output-dir <output-dir> --evaluation-dataset {librispeech-debug,librispeech,earnings22}
Install additional dependencies via:
pip install -e '.[evals,pipelines]'
By default, this command uses the latest main
branch commits from WhisperKit
and searches within Argmax-published model repositories. For optional arguments related to code and model versioning, please see the help menu with -h
We continually publish the evaluation results of Argmax-hosted models here as part of our continuous integration tests.
If you would like to evaluate WhisperKit models on your own dataset:
- Step 1: Publish a dataset on the Hub with the same simple structure as this toy dataset (audio files +
metadata.json
) - Step 2: Run evaluation with environment variables as follows:
export CUSTOM_EVAL_DATASET="my-dataset-name-on-hub"
export DATASET_REPO_OWNER="my-user-or-org-name-on-hub"
export MODEL_REPO_ID="my-org/my-whisper-repo-name" # if evaluating self-published models
whisperkit-evaluate-model --model-version <model-version> --output-dir <output-dir> --evaluation-dataset my-dataset-name-on-hub
Use the unified Python wrapper for several on-device Whisper frameworks:
Install additional dependencies via:
pip install -e '.[pipelines]'
from whisperkit.pipelines import WhisperKit, WhisperCpp, WhisperMLX, WhisperOpenAIAPI
pipe = WhisperKit(whisper_version="openai/whisper-large-v3", out_dir="/path/to/out/dir")
print(pipe("audio.{wav,flac,mp3}"))
Note: WhisperCpp
requires ffmpeg
to be installed. Recommended installation is with brew install ffmpeg
Note: WhisperOpenAIAPI
requires setting OPENAI_API_KEY
as an environment variable
This app serves two purposes:
- Base template for developers to freely customize and integrate parts into their own app
- Real-world testing/debugging utility for custom Whisper versions or WhisperKit features before/without building an app.
Note that the app is in beta and we are actively seeking feedback to improve it before widely distributing it.
Please visit the WhisperKit Benchmarks Hugging Face Space for detailed benchmark results. Here is a brief explanation to help with navigation of the results. This benchmark is updated for every non-patch release on virtually all supported devices.
Q1: xcrun: error: unable to find utility "coremlcompiler", not a developer tool or in PATH
A1: Ensure Xcode is installed on your Mac and run sudo xcode-select --switch /Applications/Xcode.app/Contents/Developer
.
If you use WhisperKit for something cool or just find it useful, please drop us a note at [email protected]!
If you use WhisperKit for academic work, here is the BibTeX:
@misc{whisperkit-argmax,
title = {WhisperKit},
author = {Argmax, Inc.},
year = {2024},
URL = {https://github.com/argmaxinc/WhisperKit}
}