Shashi Kant

Results 24 comments of Shashi Kant

Hi @colinator can you run some WER checks on medium and medium.en models for TensorRT-LLM backend? According to TensorRT-LLM repo, they only support large model. You can use these to...

> Do have a script that performs WER calculation from the csv outputs? I see your WER function, but am not totally clear on any pre-processing (lowercasing, etc) you do...

Hi @nexuslux thanks for the interest. Speaker diarization is not planned somewhere near, I may add it later on. Current priorities are: - WhisperS2T-Server - Improve TensorRT-LLM backend. - Explore...

I think whisper.cpp does not support batching. Do you know of any community implementation for batched whisper.cpp ?

Hi @joaogabrieljunq This project is more inclined toward offline ASR. Though I have some plans to work on streaming ASR using whisper in the future. Doing streaming ASR with whisper...

@OValery16 https://github.com/NVIDIA/TensorRT-LLM/issues/1127

Hi @yuekaizhang thanks for showing interest in this project! > If you're not in a hurry, I will be available to help you after 20 days. Yeah no hurry.. I...

Hey @sanchit-gandhi ! You can prepare the benchmark env using this script: [https://github.com/shashikg/WhisperS2T/blob/main/prepare_benchmark_env.sh](https://github.com/shashikg/WhisperS2T/blob/main/prepare_benchmark_env.sh). This will download the required datasets. Please also check these numbers for distil-whisper: [https://github.com/shashikg/WhisperS2T/releases/tag/v1.1.0](https://github.com/shashikg/WhisperS2T/releases/tag/v1.1.0) > Once we...

Hi @brunjo, yes, if you don't provide the language, it will fall back to English. I'll try adding that. However, the issue with this approach is that it complicates the...