S. Neuhaus
S. Neuhaus
You can also run Llama 65B (a bit slow but not terrible) on a CPU and with 128GB RAM with llama.cpp See the discussion at https://github.com/ggerganov/llama.cpp/issues/34
> > You can also run Llama 65B (a bit slow but not terrible) on a CPU and with 128GB RAM with llama.cpp > > You don't need 128GB RAM,...
OK i tried it with the fp16 model too, it only swapped a little bit (i have an 8-core Ryzen 7 3700X and 128GB RAM): ``` $ ./main -m models/65B/ggml-model-f16.bin...
As of now "quantize" only knows how to do 4bit.
I believe to have noticed a significant quality increase going from 7B to 13B and from 13B to 30B (on GPU) and i've just started with 65B and it is...
Is it swapping? Instead of opening new issues perhaps these numbers should be collected in issue #34 ("benchmarks?").
Just FYI from https://openai.com/brand > Don’t: > > Use OpenAI models in product or app names: GPT4Meow, MeowGPT I don't think OpenAI will succeed in claiming rights to the acronym...
Did you use the same email address? @sydneymwong
Thanks @edward-shen for giving the clues regarding `GRIST_SINGLE_ORG=docs` and `GRIST_ORG_IN_PATH=true`, without these options i was getting a main page not found.
This issue is still relevant.