InternVL
InternVL copied to clipboard
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型
Llava has various quantized models in gguf format, so it can be used with Llama.cpp. https://github.com/ggerganov/llama.cpp/pull/3436 Is this possible?
hi, can i extract image and text feature respectively in InternVL-G model? when read code, i found the cross-attention layers in QLLaMA are the shared parameters bewteen image and text...
Thank you for your amazing and effective work, the results of the experiment are outstanding and exciting. Do you have any plans to update the results of the model to...
LMDeploy has supported InternVL-Chat (llava, v1.2) since v0.4.0. And in v0.4.1, it adds InternVL-Chat-1.5
ERROR about flash_attr, can u help to provide version for these old nv card? --------------------------- out, q, k, v, out_padded, softmax_lse, S_dmask, rng_state = flash_attn_cuda.fwd( ^^^^^^^^^^^^^^^^^^^^ RuntimeError: FlashAttention only supports...
ValueError: Target modules ['self_attn.q_proj', 'self_attn.k_proj', 'self_attn.v_proj', 'self_attn.o_proj', 'mlp.gate_proj', 'mlp.down_proj', 'mlp.up_proj'] not found in the base model. Please check the target modules and try again. 请问有人遇到过这个问题吗?要怎么解决呢?
When I follow the method in "how_to_evaluate_internvl_chat_1_5.md" , I found that there is no such file in your repo data release. 404 page not found when run: wget https://github.com/OpenGVLab/InternVL/releases/download/data/textvqa_val.jsonl
Some weights of LlavaLlamaForCausalLM were not initialized from the model checkpoint at /data/workspace/models/InternVL-C hat-V1-5 and are newly initialized: You should probably TRAIN this model on a down-stream task to be...
Hi team, I am working on a website for a project at my school for the entrepreneurship course and I need your help. I want to integrate OpenGVLab's InternVL-Chat-V1.5 model...
Hi, Thanks for sharing the model and code with us. I am trying to using Vision Foundation Model for a zero shot classification problem. It is possible with **OpenGVLab/InternVL-14B-224px** but...