This tool extracts audio from video files and transcribes it using OpenAI's Whisper model.
- Python 3.11 or later
- FFmpeg installed on your system
- uv (Python package installer)
- CUDA-compatible GPU (optional, for faster transcription)
brew install ffmpeg
sudo apt update
sudo apt install ffmpeg
Download from ffmpeg.org and add ffmpeg
into the user PATH
- Download the
.7z
or.zip
file. - Extract the downloaded ZIP file to a folder
ffmpeg
. - Edit the path variable.
Path
->New
-> place theffmpeg/bin
directory. - Save and restart the terminal.
Before running the project, please check if ffmpeg
is working in you computer or not.
ffmpeg -version
pip install uv
Clone this repository and navigate to the project directory:
git clone https://github.com/Badhansen/transcribe.git
cd transcribe
uv sync
source .venv/bin/activate # On Windows, use `.venv\Scripts\activate`
- Place your video file in the project directory
- Update the video filename in
main.py
if needed - Run the transcription:
python main.py
The transcription will be saved to transcript.txt
.
- ffmpeg-python: Python bindings for FFmpeg
- openai-whisper: OpenAI's Whisper speech recognition model
- PyTorch: Required for Whisper model execution