llama.cpp
llama.cpp copied to clipboard
我想convert一个比较大的模型时报错Unable to allocate 1.96 GiB for an array with shape (128256, 8192) and data type float16如何解决
What happened?
模型有141G
Name and Version
latest
What operating system are you seeing the problem on?
No response
Relevant log output
No response
Looks like classic OOM (out-of-memory) error, I'd advise using smaller model or more quantized model. Without much context, it is hard to tell the issue.
This issue was closed because it has been inactive for 14 days since being marked as stale.