LLaVA
LLaVA copied to clipboard
[Question] ValueError: weight is on the meta device, we need a `value` to put in on 0.
Question
(llava) ~/autodl-tmp/LLaVA# python -m llava.serve.model_worker --host 0.0.0.0 --controller http://localhost:10000 --port 40000 --worker http://localhost:40000 --model-path liuhaotian/llava-v1-0719-336px-lora-vicuna-13b-v1.3 --model-base vicuna-13b-v1.3
[2023-10-13 17:34:01,478] [INFO] [real_accelerator.py:110:get_accelerator] Setting ds_accelerator to cuda (auto detect)
2023-10-13 17:34:01 | INFO | model_worker | args: Namespace(host='0.0.0.0', port=40000, worker_address='http://localhost:40000', controller_address='http://localhost:10000', model_path='liuhaotian/llava-v1-0719-336px-lora-vicuna-13b-v1.3', model_base='vicuna-13b-v1.3', model_name=None, device='cuda', multi_modal=False, limit_model_concurrency=5, stream_interval=1, no_register=False, load_8bit=False, load_4bit=False)
2023-10-13 17:34:01 | INFO | model_worker | Loading the model llava-v1-0719-336px-lora-vicuna-13b-v1.3 on worker 56531b ...
You are using the legacy behaviour of the <class 'transformers.models.llama.tokenization_llama.LlamaTokenizer'>. This means that tokens that come after special tokens will not be properly handled. We recommend you to read the related pull request available at https://github.com/huggingface/transformers/pull/24565
2023-10-13 17:34:03 | INFO | stdout | Loading LLaVA from base model...
Loading checkpoint shards: 0%| | 0/3 [00:00<?, ?it/s]
Loading checkpoint shards: 33%|████████████████████████████████████▎ | 1/3 [00:13<00:27, 13.68s/it]
Loading checkpoint shards: 67%|████████████████████████████████████████████████████████████████████████▋ | 2/3 [00:24<00:11, 11.77s/it]
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████| 3/3 [00:29<00:00, 8.88s/it]
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████| 3/3 [00:29<00:00, 9.85s/it]
2023-10-13 17:34:33 | ERROR | stderr |
2023-10-13 17:34:34 | ERROR | stderr | Traceback (most recent call last):
2023-10-13 17:34:34 | ERROR | stderr | File "/root/miniconda3/envs/llava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
2023-10-13 17:34:34 | ERROR | stderr | return _run_code(code, main_globals, None,
2023-10-13 17:34:34 | ERROR | stderr | File "/root/miniconda3/envs/llava/lib/python3.10/runpy.py", line 86, in _run_code
2023-10-13 17:34:34 | ERROR | stderr | exec(code, run_globals)
2023-10-13 17:34:34 | ERROR | stderr | File "/root/autodl-tmp/LLaVA/llava/serve/model_worker.py", line 275, in value
to put in on {device}.")
2023-10-13 17:34:34 | ERROR | stderr | ValueError: weight is on the meta device, we need a value
to put in on 0.
My device: GPU 0: NVIDIA GeForce RTX 3090
How can I solve this problem?
i have the same question, how to solve this problem?
same question, anyone solved it?