taozhiyuai

Results 117 comments of taozhiyuai

open as many terminals as you can

> > open as many terminals as you can > > I ended up typing this command: > > ``` > ollama pull llava-phi3; ollama pull llava-llama3; ollama pull llama3-gradient;...

is it the right model file? or I should merge them into one GGUF? FROM ./Meta-Llama-3-120B-Instruct-Q5_K_M-00001-of-00003.gguf FROM ./Meta-Llama-3-120B-Instruct-Q5_K_M-00002-of-00003.gguf FROM ./Meta-Llama-3-120B-Instruct-Q5_K_M-00003-of-00003.gguf FROM ./Meta-Llama-3-120B-Instruct.imatrix TEMPLATE """{{ if .System }}system {{ .System }}{{...

已经处理过这个问题,搜索一下

he may ask to keep model in GPU RAM all the time, even it is not in use for a very long time.

> > he may ask to keep model in GPU RAM all the time, even it does not used for a very long time. > > Yes, that's what it...

多的是中文版,搜索一下

deepseek team does not offer gguf, which in not in their plan at the moment. and its performance is not much better than llama-3-70b, but 3 times bigger in parameters...

@RexWzh @jmorganca for the same model or quantizations , is that possible the website combine them into one link. so difference users can contribute together.

Last login: Tue May 7 07:23:10 on ttys001 taozhiyu@603e5f4a42f1 Q5KM % ollama create meta-llama-3-120b-instruct:q5_k_m -f modelfile transferring model data creating model layer creating template layer creating parameters layer creating config...