Gpt4All-webui
Gpt4All-webui copied to clipboard
Inference time
Can you please let me know the inference time for the models ?
As far as i know, running gpt4all models on cpu is quite slow. i would appreciate your help in this
It highly depends on: The binding used The model size The activation of cuda
Can you point me towards one of the best GGML based model and binding ! I really need it for. Project and I am confused