Junru Shao
Junru Shao
This is weird because my 14 pro runs properly without this issue. 14 ProMax comes with 6GB memory, which means they should work smoothly without the issue you guys mentioned.
This should work now if you upgrade to the latest iOS :-)
Hey thanks for asking! MLC LLM doesn't limit which language to use per se, but good language support really depends on what underlying model it is generating code for. For...
MLC LLM is a compiler, which doesn't control the text generation from models per se. Currently the "as an AI language model" is directly generated from Vicuna-7b and we didn't...
We haven't particularly focused a lot on CUDA optimization yet, as `torch.compile()` should already work for huggingface models out of the box. Happy to bring in more optimization later
Please check out this page for building `mlc_chat_cli`: https://mlc.ai/mlc-llm/docs/tutorials/runtime/cpp.html
Please check out this page for building `mlc_chat_cli` properly from source: https://mlc.ai/mlc-llm/docs/tutorials/runtime/cpp.html
Please refer to this documentation for details: https://mlc.ai/mlc-llm/docs/tutorials/runtime/android.html
Please check out this page for proper installation of TVM: https://github.com/mlc-ai/mlc-llm/blob/main/docs/install/tvm.rst
very cool work!