Milan
Milan
Can we increase the context length say upto 4k, by fine-tuning it?
(base) volleedemo@Vollees-MacBook-Pro vollee-demo % git clone https://github.com/argmaxinc/whisperkit.git cd whisperkit Cloning into 'whisperkit'... remote: Enumerating objects: 367, done. remote: Counting objects: 100% (189/189), done. remote: Compressing objects: 100% (104/104), done. remote:...
Tried with all available tags too, same error.
swift run transcribe --model-path "Models/whisperkit-coreml/openai_whisper-large-v3" --stream --verbose Fetching https://github.com/huggingface/swift-transformers.git from cache Fetching https://github.com/apple/swift-argument-parser.git from cache Fetched https://github.com/apple/swift-argument-parser.git (1.19s) Fetched https://github.com/huggingface/swift-transformers.git (1.19s) Computing version for https://github.com/apple/swift-argument-parser.git Computed https://github.com/apple/swift-argument-parser.git at 1.3.0 (0.03s)...
> Thanks @milsun, could you also share your HW details? Is this an Apple Silicon Mac? My first comment has all the HW details.
Can you share the specific versions, for which it works, need to get it in place for a project I have been working on. Thanks!
Did a fresh clone and install, still the same issue.
this is a really cool feature in my opinion, any luck so far?
would denoising or echo cancellation type approach really solve this issue?
> > If you need Diarization, migrate to RevAI's reverb model. > > I came back to this repo after reading how RevAI's Reverb is apparently the new SOTA English...