Robin
Robin
Here some other automl-toolkit have supported spark such as Hyperopt and Optuna, maybe we can refer and discuss how to support MLLib in NNI?
试一下用微软的 deepspeed 呢
是使用的 lora 训练的吗
Hi, NielsRogge, I am also very confused when using the code. It seems like he did not process the dataset in a conversational manner.
Yes, 3\*8*A100 (40G) is enough for fine-tuning llama-2-70B
Here is the merge code: https://github.com/jianzhnie/Efficient-Tuning-LLMs/blob/main/chatllms/utils/apply_lora.py
https://github.com/jianzhnie/Efficient-Tuning-LLMs/blob/main/chatllms/utils/apply_lora.py
CUDA_VISIBLE_DEVICES=0 python chatllms/utils/apply_lora.py \ --base-model-path ~/checkpoints/baichuan7b/ \ --lora-model-path ./work_dir/vicuna_merge_vicuna-baichuan-7b-1gpu/checkpoint-15000 \ --target-model-path ./work_dir/vicuna_merge_vicuna-baichuan-7b/merged_model