sweetcard

Results 30 comments of sweetcard

The tokenizer by pytorch has poor performance and uses too much memory. Replacing it with mlx will be good news.😄

ENV CATALINA_OPTS="-Dorg.apache.tomcat.util.buf.UDecoder.ALLOW_ENCODED_SLASH=true" ENV JAVA_OPTS="-Duser.language=zh -Duser.region=CN -Dfile.encoding=UTF-8" Set env for tomcat and it should work.

Maybe LM studio doesn't update to the latest version of Llama.cpp. Be patient and wait for some time😄

> Thank you for interest in MiniCPM. I am one of the authors. In MiniCPM, we implement `tie_word_embedding`, which involves utilizing the same matrix for both input embedding and the...

> > > Thank you for interest in MiniCPM. I am one of the authors. In MiniCPM, we implement `tie_word_embedding`, which involves utilizing the same matrix for both input embedding...

> > > > > Thank you for interest in MiniCPM. I am one of the authors. In MiniCPM, we implement `tie_word_embedding`, which involves utilizing the same matrix for both...

> A good news is that we have converted the original checkpoints into Llama format. Specifically, > > 1. we absorb the $mu$P scaling factors into the model checkpoints. >...

用llama.cpp吧。都有了。😄

可以试试ollama,一条命令即可体验。