Zhihong Chen

Results 31 comments of Zhihong Chen

Hi @intelligencegear and @RahulBhalley, Thanks for your attention! We have uploaded the training code (see [here](https://github.com/FreedomIntelligence/LLMZoo#-training-by-yourself)). Now you can train `Phoenix`. :-) We specify all the hyper-parameters in the shell...

Hi @ananwjq, Thanks for your feedback! We are working on efficient inference and will give a detailed comparison. Best, Zhihong

Hi @yuyq96, Thanks for your attention! Could you upgrade the `transformers` version: `pip install git+https://github.com/huggingface/transformers` and try again? Best, Zhihong

Hi all, Please check if the dataset is the same one used in the codebase. Here is the [link](https://drive.google.com/file/d/1QU7OFll-vpB7UNLySoyUJbxVJfc_Y9zS/view?usp=sharing). Best, Zhihong

Hi @feiyacz, 可以提供一下更具体的信息吗,比如其中哪个库安装失败。 Best, Zhihong

Hi @YerayL, Could you provide the `transforemers` version? Best, Zhihong

Dear @LiZhangMing, Thanks! For the issue about `Flast-Attn`, please refer to [this repo](https://github.com/HazyResearch/flash-attention). Alternatively, we can use `train.py` rather than `train_fast.py` to turn off the flash attention. Best, Zhihong

Hi @mohataher, Thanks! We did not extend the vocabulary size for llama. Best, Zhihong

Hi @Maydaytyh, Thanks for your attention! Set `--nproc_per_node` to the number of GPUs, i.e., `2`. Best, Zhihong

Hi @porridgeshoes, 单卡目前还跑不起来,需要用更多的一些技巧,可以参考现有的[仓库](https://github.com/artidoro/qlora). Best, Zhihong