VibeVoice
VibeVoice copied to clipboard
Open-Source Frontier Voice AI
**Description:** Currently, generated audio using VibeVoice tends to be continuous, which can sound unnatural, especially for longer content like lectures or podcasts. It would be extremely helpful to have the...
I am having this issue but am not able to understand exactly how to fix it using this suggestion on the github page - 'If you found the generated voice...
Would there be interest in modifying the demo code `demo/inference_from_file.py` to support [Intel XPUs](https://docs.pytorch.org/docs/stable/notes/get_start_xpu.html)? If so, I have it working [here](https://git.ayo.run/ayo/VibeVoice/commit/1cb2a50ce5954d5871e2556f6a97a2be81cdcf9c) and would be happy to open a PR. Thanks!...
python demo/inference_from_file.py --model_path microsoft/VibeVoice-Large --txt_path demo/text_examples/2p_music.txt --speaker_names Alice Frank
I really want it as my tts server.
when I run this: python demo/inference_from_file.py --model_path microsoft/VibeVoice-1.5B --txt_path demo/text_examples/1p_abs.txt --speaker_names Alice I get some error, I find it occurs at VibeVoiceForConditionalGenerationInference/forward/logits = self.lm_head(hidden_states[:, slice_indices, :]) I print "hidden_states[:, slice_indices,...
I’m running VibeVoice on an AMD GPU with torch-directml as the backend, and I hit a runtime error during inference: ```bash \vibevoice\modular\modular_vibevoice_tokenizer.py", line 495, in _forward_streaming full_input = torch.cat([cached_input, x],...
Here you can see how it goes from generate to stop, then flashes generate long enough for you to want to click on it but then replaces it with "random...
Hello, do you have a Chinese voice tone model?