llama.cpp
llama.cpp copied to clipboard
Does llama.cpp deploy support mutil_nodes mutil-GPUs
I have two machine with 2 * 8 * A800, want deploy a GGUF model with two machines。 Does llama.cpp deploy support mutil_nodes mutil-GPUs , if OK, How can I do this ?