Tungsong

Results 6 issues of Tungsong

**I trained Alpaca-LoRA model with params:** base_model: /usr/local/dbbd/model/llama-7b-hf data_path: alpaca_data.json output_dir: ./lora-alpaca batch_size: 128 micro_batch_size: 4 num_epochs: 2 learning_rate: 0.0001 cutoff_len: 512 val_set_size: 2000 lora_r: 8 lora_alpha: 16 lora_dropout: 0.05...

I have two T4 on my machine, and I want to improve training efficiency, because it has enough memory when I use the default params ![企业微信截图_16814425311604](https://user-images.githubusercontent.com/35001280/231933309-606a555a-b870-4c9b-b864-c51320895cbb.png) I tried to update...

微调完给出的answer一直在重复同一句话,而且也不是答案 微调方式参考的https://github.com/27182812/ChatGLM-LLaMA-chinese-insturct 训练了3个epoch,效果如下,请教有谁知道为什么会这样吗 ![企业微信截图_16810901118715](https://user-images.githubusercontent.com/35001280/230808375-697f6a51-2b36-4daf-bfc3-69d5518e994c.png)

N次循环会初始化N次还是只初始化一次,怎么初始化的呢,不太懂这个α怎么来的,有什么作用 ![企业微信截图_16904561135872](https://github.com/PKU-YuanGroup/ChatLaw/assets/35001280/c99f2fd9-9955-4738-987e-f198b4ba7b8d)

多卡推理参考的ChatGLM-6B官方的多卡部署 https://github.com/THUDM/ChatGLM-6B#%E5%A4%9A%E5%8D%A1%E9%83%A8%E7%BD%B2

不懂这行代码有什么作用,引入时报错 ![image](https://github.com/FlagOpen/FlagEmbedding/assets/35001280/7c79bf83-5715-4387-8ca8-3ec70a403305)