Ziqing Yang

Results 212 comments of Ziqing Yang

1. Chinese-LLaMA可以使用float16配合peft进行LoRA训练,训练脚本里也默认开启了fp16 2. 暂无明确计划,或许会考虑针对个别场景采用更长的block size

参考 https://github.com/ymcui/Chinese-LLaMA-Alpaca/issues/361

Would you mind if I copy and put your Colab notebook into the repo, so that the repo is self-contained?

After I click the submit, html tags are added to the texts (The input is "你好"): Is there a way to remove them? @sunyuhan19981208

@sunyuhan19981208 My colleague and I can both reproduce the problem with the notebook link you gave https://colab.research.google.com/drive/1mG8CwGub-hcoeo79FjfoWtxqOuu3P_y1?usp=sharing. We have tested Safari + MacOS, Chrome + MacOS, Chrome + Win11, Edge...

> > @sunyuhan19981208 My colleague and I can both reproduce the problem with the notebook link you gave https://colab.research.google.com/drive/1mG8CwGub-hcoeo79FjfoWtxqOuu3P_y1?usp=sharing. > > We have tested Safari + MacOS, Chrome + MacOS,...

We temporarily merge the PR into `gradio_notebook` branch, and then merge it into the main branch after reorganizing some code and docs.

1. 可以在新的合适的数据上继续精调模型,提升降低拒答概率 2. CPU精调目前实践上基本不太可行,见 #346

`data/pt_data.txt`只是示例,预训练数据与指令数据无关, 比如可以采用中文wiki语料,悟道语料(WuDaoCorpus)、ROOTS数据集中的中文部分等语料

> 好的, > > 几个问题哈, > > 1)那我是在合并之后的模型基础上微调的,是不是应该和合并后的模型再合并,而不是去和原版llama-hf合并? 2)合并还是用merge_llama_with_chinese_lora.py么 1) 对,你基于什么模型训练,就和什么模型去合并 2) 可以用merge_llama_with_chinese_lora.py,但需要准备好相关模型权重和配置文件,仿照我们发布的LoRA的内容即可。