Results 130 comments of ldwang

先关闭,如有问题重新打开issue,谢谢

先关闭,如有问题重新打开issue,谢谢

先关闭了,如有需要重新打开。谢谢

you can manually download model by visiting https://model.baai.ac.cn/models and searching by Aquila. By running example, models will downloaded automatically. huggingface usage is coming soon.

可以修改下 checkpoints_in/aquilachat-7b/config.json "flash_atten": true 修改成 "flash_atten": false

或者删除 checkpoints_in/aquilachat-7b/config.json 重试下

后续可以支持xformers。使用flash_atten主要因为训练效率。可以修改模型配置中关闭flash_atten,使用原始atten来推理。

`conv = default_conversation.copy() conv.append_message(conv.roles[0], human_text) conv.append_message(conv.roles[1], bot_text) conv.append_message(conv.roles[0], text) conv.append_message(conv.roles[1], None)` 多轮的语料比较少。prompt构造可以这么试试,https://github.com/FlagAI-Open/FlagAI/blob/master/examples/Aquila/Aquila-chat/generate_chat.py#L39

预计发布1.7.2 会增加多轮的例子。

Aquila预训练使用了Pile,[RedPajama-Data-1T](https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T), [Wikipedia](https://huggingface.co/datasets/wikipedia), [C4](https://huggingface.co/datasets/c4), 悟道中文数据集、电子书、专利、百科、论坛, github数据等