CanCanNeed
CanCanNeed
### Question Hello, have you open sourced the code for comparative experiments using the Qwen-VL model?
### Question I am using openai's visual encoder clip-vit-large-patch14-336, but a problem occurred: ValueError: Unknown image tower: /hy-tmp/LLaVA/clip-vit-large-patch14-336. please tell me how to solve this problem
你好,我对你的工作非常感兴趣。但是我在运行代码的时候出现了一个问题。我使用了一块48G的A6000显卡,CUDA版本是12.2,请问您有解决思路吗?谢谢。 Traceback (most recent call last): File "/hy-tmp/VSA/app.py", line 102, in vsa = VisionSearchAssistant() File "/hy-tmp/VSA/models/vsa_model.py", line 298, in __init__ self.vlm = VLM( File "/hy-tmp/VSA/models/vsa_model.py", line 132, in __init__ self.tokenizer,...