yukang
yukang
Yes. 0.5 is the right setting. Would you please show me how you calculate FLOPs? I used the method provided in this work. https://github.com/CVMI-Lab/SparseKD
Sorry for the late reply. You can find them here https://github.com/NVlabs/VILA/tree/main/longvila
Thanks for your interests in our work. We are waiting for legal permission from NVIDIA to release qwen2 models. It should be approved this week. I will let you know...
@liyucheng09 Hi, our code and models have been released. You can find it below. The benchmark comparison uses Efficient-Large-Model/qwen2-7b-longvila-256f. Updated results is in the paper. Paper: https://arxiv.org/pdf/2408.10188 Code: https://github.com/NVlabs/VILA/tree/main/longvila Model:...
Hi @liyucheng09 , Sorry for the bug. I fix it in this pr. You can try this repo, https://github.com/yukang2017/VILA/tree/main, before it was merged. https://github.com/NVlabs/VILA/pull/170 Regards, Yukang Chen
Hi @liyucheng09 , I think we do not need to set conv_template when use vila-run evaluation. It can find the conv template from the tokenizer config json. I did not...
Hi, I set `inputs_embeds` as `(num_tokens, embed_dim)` and get the following issues. Are there any advice? Thanks. 
@qthequartermasterman Thanks for your help. The index range error is gone. But the outputted text is still gibberish.