Remc
Remc
Why did you use whisperX for loading? Is there method especially efficient?
I am working on the same subject, you can find work done by Majdoddin here: https://github.com/Majdoddin/nlp Not perfect but a good way to start. I ll push my solution when...
If it happen, it will be link to faster_whisper implementation and not on whisperX side. Good news, their is an issue open on faster_whisper repo : https://github.com/guillaumekln/faster-whisper/issues/533
You can finetune faster_whisper models. Just fine tune a regular whisper model and you will be able to pass those weigth to faster_whisper model.
When using int8 on Mac with m2, whisperx responds `TypeError: TranscriptionOptions.__new__() missing 3 required positional arguments: 'max_new_tokens', 'clip_timestamps', and 'hallucination_silence_threshold'`