Made in Vancouver, Canada by Picovoice
The purpose of this benchmarking framework is to provide a scientific comparison between different voice activity engines in terms of accuracy metrics. While working on Cobra, we noted that there is a need for such a tool to empower customers to make data-driven decisions.
LibriSpeech (test_clean portion) is used as the voice dataset. It can be downloaded from OpenSLR.
In order to simulate real-world situations, the data is mixed with noise (at 0dB SNR). For this purpose, we use the DEMAND dataset, which has noise recording in 18 different environments (e.g. kitchen, office, traffic, etc.). Recordings containing distinct voice data are filtered out. It can be downloaded from Kaggle.
Three voice-activity engines are used:
- py-webrtcvad (Python bindings to the WEBRTC VAD), which can be installed using PyPI.
- Cobra, which is included as submodules in this repository.
- Silero VAD, which can be installed using PyPI. Version 5.1.
We measured the accuracy of the voice activity engines using false positive and true positive rates. The false positive rate is measured as the number of false positive frames detected over the total number of non-voice frames. Likewise, true positive rate is measured as the number of true positive frames detected over the total number of voice-frames. Using these definitions, we plot a receiver operating characteristic curve which can be used to characterize performance differences between engines.
The benchmark has been developed on Ubuntu 18.04 with Python 3.8. Clone the repository using
git clone https://github.com/Picovoice/voice-activity-benchmark.gitMake sure the Python packages in the requirements.txt are properly installed for your Python version as Python bindings are used for running the engines.
Usage information can be retrieved via
python benchmark.py -hThe runtime benchmark for C is contained in the runtime folder. Use the following commands to build and run the runtime benchmark:
git clone --recursive https://github.com/Picovoice/cobra.git runtime/cobra
cmake -S runtime -B runtime/build && cmake --build runtime/build
./runtime/build/cobra_runtime -l {COBRA_LIBRARY_PATH} -a {ACCESS_KEY} -w {TEST_WAVFILE_PATH}The runtime benchmark for Python can be run with the following command:
python3 -m runtime.runtime_test --wav_path {TEST_WAVFILE_PATH} --engine {"Cobra" | "Silero" | "WebRTC"} --access_key {ACCESS_KEY}Below is the result of running the benchmark framework. The plots below shows the receiver operating characteristic curve of different engines. The top shows the overall graph. The bottom shows the top-left corner with a minimum true-positive rate of 50% and a maximum false-positive rate of 5%. The plots were generated with a signal-to-noise ratio of 0dB.
On a Raspberry Pi Zero, Cobra measured a realtime factor of 0.05, or about 5% CPU usage.
On a laptop with an Intel(R) Core(TM) i7-1185G7, Cobra measured a realtime factor of 0.0006.
On an Ubuntu 22.04 machine with AMD CPU (AMD Ryzen 9 5900X (12) @ 3.70GHz), the following RTFs were recorded:


