Robin

Results 36 comments of Robin

Here some other automl-toolkit have supported spark such as Hyperopt and Optuna, maybe we can refer and discuss how to support MLLib in NNI?

试一下用微软的 deepspeed 呢

是使用的 lora 训练的吗

Hi, NielsRogge, I am also very confused when using the code. It seems like he did not process the dataset in a conversational manner.

Yes, 3\*8*A100 (40G) is enough for fine-tuning llama-2-70B

Here is the merge code: https://github.com/jianzhnie/Efficient-Tuning-LLMs/blob/main/chatllms/utils/apply_lora.py

https://github.com/jianzhnie/Efficient-Tuning-LLMs/blob/main/chatllms/utils/apply_lora.py

CUDA_VISIBLE_DEVICES=0 python chatllms/utils/apply_lora.py \ --base-model-path ~/checkpoints/baichuan7b/ \ --lora-model-path ./work_dir/vicuna_merge_vicuna-baichuan-7b-1gpu/checkpoint-15000 \ --target-model-path ./work_dir/vicuna_merge_vicuna-baichuan-7b/merged_model