Zhengxiao Du

Results 163 comments of Zhengxiao Du

这个问题应该是新版本的DeepSpeed将`_restore_from_fp16_weights`重命名为了`_restore_from_bit16_weights`,但同时在DeepSpeedEngine中没有修改。一种解决方法是使用旧版本的DeepSpeed(我们使用的是0.3.16),另一种是修改deepspeed/runtime/engine.py,将`self.optimizer._restore_from_fp16_weights()`这行修改为`self.optimizer._restore_from_bit16_weights()`

应该是{"prompt": "", text: 小说文本}。prompt表示的是不需要生成的部分,如果你是想生成小说的话可以从头生成没有prompt。按512长度分段是程序自动完成的,不需要在数据里面完成

The two ways of multi-task learning are designed for ablation study of different objectives. To enable adaption to different downstream tasks, we mix all the three types of objectives in...

These are arguments of the DeepSpeed launcher. `NUM_WORKERS` is used to set `--num_nodes`, which means the number of servers used for pretraining `NUM_GPUS_PER_WORKER` is used to set `--num_gpus`, which means...

You need to divide the downloaded checkpoint with `change_mp.py`, following the instruction in https://github.com/THUDM/GLM#model-parallelism

You cannot install apex with `pip install apex` because there is another package apex that has nothing to do with deep learning in PyPI, which is why we didn't list...

> Any update on this? I notice that there is a out of box pretrain version for GLM-10B. Would like to know whether there are any future plan on uploading...

你手动运行一下 `ctypes.cdll.LoadLibrary("C:\Users\oo.cache\huggingface\modules\transformers_modules\chatglm-6b-int4\quantization_kernels.so")` 看一下会报什么错?

可以 pull 一下最新的仓库代码

这个需求可以直接用量化后的模型 https://huggingface.co/THUDM/chatglm-6b-int4 和 https://huggingface.co/THUDM/chatglm-6b-int8 不过量化之后在 GPU 上推理也是需要用 CUDA写的 kernel的,我觉得可能无法成功。要解决这个问题还是要把 CUDA kernel 移植到 ROCm