mlc-llm
mlc-llm copied to clipboard
[Question] Whether to support running on ARM64v8
❓ General Questions
When I completed the compilation and tried to run ./mlc_chat_cli --local-id ../../dist/vicuna-7b-all-v1.1-q3f16_0
, I got follows error:
`
terminate called after throwing an instance of 'tvm::runtime::InternalError'
what(): [06:45:08] /data/lxxin/mlc-llm/cpp/cli_main.cc:50: Cannot auto detect device-name
Stack trace:
[bt] (0) /data/ps/mlc-llm/build/tvm/libtvm_runtime.so(tvm::runtime::Backtraceabi:cxx11+0x1c) [0xffff76d377ac]
[bt] (1) ./mlc_chat_cli(tvm::runtime::detail::LogFatal::Entry::Finalize()+0x54) [0xaaaac9aaee74]
[bt] (2) ./mlc_chat_cli(+0x8cb4) [0xaaaac9aa8cb4]
[bt] (3) ./mlc_chat_cli(+0x8ed4) [0xaaaac9aa8ed4]
[bt] (4) ./mlc_chat_cli(+0x76b4) [0xaaaac9aa76b4]
[bt] (5) /lib/aarch64-linux-gnu/libc.so.6(__libc_start_main+0xe8) [0xffff76890e18]
[bt] (6) ./mlc_chat_cli(+0x8128) [0xaaaac9aa8128]
Aborted (core dumped) `
Please see the documents here, need to run gen_cmake_config.py
to detect device
Much thanks! I tried it, but I still got the same error, I suppose MLC could not support running in ARM CPU(or not only ARM) without GPU.
You are right. The CPU is too weak to run LLMs, so we only focus on the GPU environment due to limited bandwidth.
On the other hand, it's not hard to add CPU if you are familiar with TVM. Happy to help if you are interested in contributing CPU support to the community.
Thanks very much! I got it.