Junru Shao

Results 179 comments of Junru Shao

This is weird because my 14 pro runs properly without this issue. 14 ProMax comes with 6GB memory, which means they should work smoothly without the issue you guys mentioned.

This should work now if you upgrade to the latest iOS :-)

Hey thanks for asking! MLC LLM doesn't limit which language to use per se, but good language support really depends on what underlying model it is generating code for. For...

MLC LLM is a compiler, which doesn't control the text generation from models per se. Currently the "as an AI language model" is directly generated from Vicuna-7b and we didn't...

We haven't particularly focused a lot on CUDA optimization yet, as `torch.compile()` should already work for huggingface models out of the box. Happy to bring in more optimization later

Please check out this page for building `mlc_chat_cli`: https://mlc.ai/mlc-llm/docs/tutorials/runtime/cpp.html

Please check out this page for building `mlc_chat_cli` properly from source: https://mlc.ai/mlc-llm/docs/tutorials/runtime/cpp.html

Please refer to this documentation for details: https://mlc.ai/mlc-llm/docs/tutorials/runtime/android.html

Please check out this page for proper installation of TVM: https://github.com/mlc-ai/mlc-llm/blob/main/docs/install/tvm.rst