Yushi Bai
Yushi Bai
如果条件允许的话,可以用多gpu推理,只需要在load模型时传入`device_map="auto"`
Thanks! Here is the English version: The training was successful! Here are the details of the environment: - Environment: - python==3.11.9 - transformers==4.33.0 - pytorch==2.2.0 - flash-attn==2.6.3 - ninja==1.11.1.1 -...
> @LYCnight 大佬 训练的时候预计要多大的显存 GLM-4-9b 32k训练需要8卡80G。如果显存不够可以试试lora或者qlora。
> @LYCnight 按照大佬的环境,成功的开启了训练,但是为什么训练完后的文件超级大呢?存储空间不够了…… -rw-r--r-- 1 root root 4984147224 Sep 5 10:49 model-00001-of-00004.safetensors -rw-r--r-- 1 root root 4895071360 Sep 5 10:49 model-00002-of-00004.safetensors -rw-r--r-- 1 root root 4895071384 Sep 5 10:49 model-00003-of-00004.safetensors...
Hi! You can get the token id by `tokenizer.get_command("")`.
Have you updated to our most recent model files? Also, please use `transformers>=4.43.0`.
This is correct. Thanks for sharing!
This is just a warning, right? You can ignore it.
你好,目前我们还没有确定的上架时间,之后我们会release具有更强的长输出性能而且更大的模型,应该会在1-2个月内。
sorry~我们暂时还不能透露