Zhe Chen

Results 316 comments of Zhe Chen

> 能请问一下什么时候开源这部分的代码吗? 您好,感谢您的关注。目前我们暂时没有计划开源从零训练ViT的代码。

Hello, the pre-training weights from the first stage are essentially the MLP projector, and we will release them shortly. Additionally, the data format for our pre-training is consistent with the...

Hello, we are planning to release some pre-trained OCR data, but the dataset is quite large, consisting of tens of millions of entries, so it will take some time to...

这个bug看着好像可以通过调整一下权重的device来解决

感谢提问,InternVL2用的ViT还是InternVL1.5的ViT,就是这个: https://huggingface.co/OpenGVLab/InternViT-6B-448px-V1-5 我们在InternVL2这个版本中没有对ViT做额外的增量预训练。

Thank you for your suggestion. We will implement the `apply_chat_template` function in the next few days.

这还挺奇怪,我没遇到过这个问题。请问您的环境中用安装apex和flasn attn吗。如果有安装apex,建议卸载了;如果没安装flash attn可以安装一个试试看,因为lmdeploy中运行vlm的ViT部分应该还是跑的pytorch后端。

We plan to complete the integration of ollama within October, thank you for waiting.

Yes, when using V100 GPUs, you can manually disable flash attention by setting `use_flash_attn=False`.