Skip to content

Releases: m-bain/whisperX

v3.3.0

02 Jan 13:09
Compare
Choose a tag to compare

What's Changed

  • Update faster-whisper to 1.0.2 to enable model distil-large-v3 by @moritzbrantner in #814
  • latest faster-whisper support added by @Hasan-Naseer in #875
  • Working version with pyannote:3.3.2 and faster-whisper:1.1.0 by @ibombonato in #936
  • Add ultization to verbose flag by @H4CK3Rabhi in #759
  • Added local_files_only option on whisperx.load_model for offline mode by @RoqueGio in #867
  • adding cache_dir to wav2vec2 by @bnitsan in #681
  • feat: add basic installation test flow & restrict python versions by @Barabazs in #965
  • chore: add build and release workflow by @Barabazs in #966
  • fix: update README image source and enhance setup.py for long description by @Barabazs in #968
  • docs: update installation instructions in README by @Barabazs in #969
  • fix: add UTF-8 encoding when reading README.md by @xigh in #970
  • chore: loosen ctranslate2 version restriction & bump whisperX version by @Barabazs in #971

New Contributors

Full Changelog: v3.2.0...v3.3.0

v3.2.0

18 Dec 08:03
Compare
Choose a tag to compare

Device and Language Support

Bug Fixes and Stability Improvements

Documentation Updates

Miscellaneous Changes

Full Changelog: v3.1.1...v3.2.0

3.1.1

13 May 11:19
d8a2b4f
Compare
Choose a tag to compare
  • translate functionality added
  • fix word timestamp bug (words no longer have consecutive timestamps)

v3.1.0

07 May 19:34
1caddfb
Compare
Choose a tag to compare
  • 70x real time transcription, <8GB gpu memory requirement ⚡️⚡️
  • each transcript segment is a sentence (using nltk.sent_tokenize)
  • diarization now assigned per sentence (and outputted to srt)
  • clean up on alignment logic

v3.0.2

04 May 19:53
Compare
Choose a tag to compare
v3.0.2 Pre-release
Pre-release

torch2.0, python3.10

v3.0.1

01 May 10:52
e24ca9e
Compare
Choose a tag to compare
v3.0.1 Pre-release
Pre-release
  • fix pickling error (set num_workers=0) to perform dataloading iwth main process
  • add basic diarization
  • pad language detection if less than 30s

v3.0.0

28 Apr 15:48
cc7e168
Compare
Choose a tag to compare
v3.0.0 Pre-release
Pre-release

batched inference with faster-whisper backend

v2.0.1

28 Apr 15:47
Compare
Choose a tag to compare
allow custom model_dir for torchaudio models

alpha

26 Feb 21:05
847a3cd
Compare
Choose a tag to compare
Merge pull request #96 from smly/fix-batch-processing

FIX: Assertion error in batch processing