xianghuisun

Results 68 comments of xianghuisun

可参考: https://github.com/LianjiaTech/BELLE/issues/91 https://github.com/LianjiaTech/BELLE/issues/122

> > > > 谢谢您的答复,我还有一个问题是能使用量化后的模型继续做finetune么??? 量化后的模型理论上是可以继续finetune的。不过目前我们并没有做相关实验

> _No description provided._ 详见train/FAQ.md 如果是全量参数微调,需要8张A100 40G 如果是LoRA,单张A100 40G

> igscience/bloomz-7b1-mt", "data_path": "data/res/merge_data.json", "output_dir": "trained_models/bloom", "per_device_train_batch_size": 1, "num_epochs": 2, "learning_rate": 1e-5, "cutoff_len": 1024, "val_set_size": 1000, "val_set_rate": 0.1, "save_steps": 1000, "eval_steps": 1000, "logging_steps": 1, "gradient_accumulation_steps": 32 } > > deepspeed...

我们会找时间尝试一下,看看能不能复现这个问题。 ------------------ 原始邮件 ------------------ 发件人: "LianjiaTech/BELLE" ***@***.***>; 发送时间: 2023年4月9日(星期天) 晚上7:08 ***@***.***>; ***@***.******@***.***>; 主题: Re: [LianjiaTech/BELLE] 出现如下warning: tried to get lr value before scheduler/optimizer started stepping, returning lr=0 (Issue #134) igscience/bloomz-7b1-mt", "data_path": "data/res/merge_data.json", "output_dir": "trained_models/bloom",...