Video-LLaVA
Video-LLaVA copied to clipboard
TypeError: 'NoneType' object is not callable
HelloοΌ@LinB203 I used the βInference for videoβ code from readme,but got
Loading checkpoint shards: 100%|ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ| 2/2 [00:06<00:00, 3.17s/it] Some weights of the model checkpoint at model were not used when initializing LlavaLlamaForCausalLM: ['model.image_tower.image_tower.encoder.layers.15.self_attn.k_proj.weight', 'model.video_tower.video_tower.encoder.layers.15.mlp.fc1.weight', 'model.image_tower.image_tower.encoder.layers.17.layer_norm1.bias', 'model.image_tower.image_tower.encoder.layers.7.layer_norm2.weight', 'model.video_tower.video_tower.encoder.layers.22.self_attn.q_proj.weight', 'model.video_tower.video_tower.encoder.layers.10.self_attn.k_proj.weight', 'model.image_tower.image_tower.encoder.layers.11.self_attn.k_proj.weight', 'model.video_tower.video_tower.encoder.layers.9.self_attn.q_proj.weight', 'model.video_tower.video_tower.encoder.layers.11.temporal_attn.v_proj.bias', γγγγγγγγγγγγγγγγγγγγγγγγγγγγγγγγγγγγγγγ 'model.video_tower.video_tower.encoder.layers.23.mlp.fc1.bias', 'model.image_tower.image_tower.encoder.layers.13.layer_norm1.weight', 'model.video_tower.video_tower.encoder.layers.11.layer_norm1.bias']
- This IS expected if you are initializing LlavaLlamaForCausalLM from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
- This IS NOT expected if you are initializing LlavaLlamaForCausalLM from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).
Traceback (most recent call last):
File "/home/win11/githubPublic/text-generation-webui/installer_files/conda/envs/videollava/lib/python3.10/runpy.py", line 196, in _run_module_as_main
return _run_code(code, main_globals, None,
File "/home/win11/githubPublic/text-generation-webui/installer_files/conda/envs/videollava/lib/python3.10/runpy.py", line 86, in _run_code
exec(code, run_globals)
File "/home/win11/Documents/Video-LLaVA/videollava/serve/cli.py", line 135, in
main(args) File "/home/win11/Documents/Video-LLaVA/videollava/serve/cli.py", line 62, in main file = video_processor(file, return_tensors='pt')['pixel_values'][0].to(model.device, dtype=torch.float16) TypeError: 'NoneType' object is not callable
It seems to have read the model correctly, But when exec tokenizer, model, processor, _ = load_pretrained_model(model_path, None, model_name, load_8bit, When load_4bit,device=device, cache_dir=cache_dir), processor returns None
Did you happen to fix this? I am actually having the same issue.
Did you happen to fix this? I am actually having the same issue.
no
I think if you put the word "llava" in your model checkpoint name it should work, I tried debugging it and that was the only solution that worked for me.