Zhe Chen

Results 316 comments of Zhe Chen

感谢反馈,我们正在解决这个问题。

> > 感谢反馈,我们正在解决这个问题。 > > 您好,请问运行vit-6b最少需要多少显存呢,需要A100 80g才可以运行吗,想单独运行来做ocr 如果是bf16的话,需要16G+的显存,int8大概8G+;不过单独用vit还做不了ocr,需要额外接上ocr的识别头。

Hi, we may release the annotation files in the JSONL format that we use. However, to make them usable for everyone, we will need to create a document detailing the...

It is indeed a bit strange; the model deployed in the demo and the open-source model have the same weights. I'll check for the reason.

I suspect a problem with device_map='auto', are you now running this model distributed across multiple GPUs?

the default `generation_config` used in the online web demo is: ```python generation_config = dict( num_beams=1, max_new_tokens=1024, do_sample=True, temperature=0.2, top_p=0.7 ) ```

Hello, thank you for your interest. We have successfully implemented the 8-bit version; however, we have encountered some issues with 4-bit quantization. We will resolve these as soon as possible...

> @czczup our team LMDeploy is working on quantizing VLMs into 4bit by AWQ The relative PR is [InternLM/lmdeploy#1553](https://github.com/InternLM/lmdeploy/pull/1553) Can we collaborate on it? That sounds great! I'm thrilled at...

The 4-bit version of the model has been released. Check it out at [OpenGVLab/InternVL-Chat-V1-5-AWQ](https://huggingface.co/OpenGVLab/InternVL-Chat-V1-5-AWQ). Thanks to the [lmdeploy](https://github.com/InternLM/lmdeploy) team for their support with model quantization. I'm closing this issue now,...

@irexyc 感谢lmdeploy大佬的支持