Yuxuan Zhang
Yuxuan Zhang
先用transformers 4.40 老模型维护少,GLM-4会跟进维护到4.45
这个bug修了,在huggingface中文件更新了
There is a big problem, your prompt is too short. Please carefully read our readme. We need to use long prompts as input, which requires you to use large language...
我觉得在这里或许能帮助到你查看代码 https://huggingface.co/THUDM/glm-4v-9b/blob/main/visual.py 视觉部分是EVA2CLIPModel,然后encoder采用的LLM模型是什么?这个问题没理解,这是个decoder only模型
36g是GPU的,没有测试过NPU的这是消耗了多少G了
需要看一下后台的报错。只有一个前端无法debug
你i应该是用了FP16吧,这个模型仅支持BF16,FP16会出现CUDA错误
This project is based on the diffusers version; we are about to launch it soon.
We have released the version of diffusers, now you can give it a try, please download the diffusers library from the source code and run the diffusers model we open-sourced...