LMFlow
LMFlow copied to clipboard
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
Thanks for creating such an amazing project. Vicuna-13b-v1.1 seems to achieve impressive performance in tasks in multiple languages. I believe it would be a good idea to conduct further fine-tuning...
Hi, thanks for this nice repo! I'm always facing kill signal and error code -11 after executing example/finetune.py with my own 10k dataset (text_only) on a single A100 40GB, CPU...
I ran the ./scripts/run_evaluation_with_lora.sh with my testing data. However, the string length of the model output seems to have a limit (Some strings in the testing data is very long),...
运行环境是colab,gpu是tsla T4,torch版本是2.0,CUDA版本是11.8,训练模型的时候出现这个报错:ValueError: Your setup doesn't support bf16/gpu. You need torch>=1.10, using Ampere GPU with cuda>=11.0
+ deepspeed --master_port=11000 examples/finetune.py --model_name_or_path /workspace/work/LMFlow/LMFlow/mydata/model/llama-7b-hf --save_aggregated_lora 0 --use_lora 1 --lora_r 8 --dataset_path /workspace/work/LMFlow/LMFlow/mydata/data/wiki_cn --block_size 512 --validation_split_percentage 0 --dataloader_num_workers 1 --output_dir /workspace/work/LMFlow/LMFlow/mydata/model/output_models/7b-wiki --overwrite_output_dir --num_train_epochs 0.01 --learning_rate 1e-4 --per_device_train_batch_size 4 --deepspeed...
我以为是设备资源的问题,因此也尝试将模型换成`git-2 samll`,也出现相同的错误。 这个模型是是来自`https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads&search=gpt2-small-chinese-cluecorpussmall`的`gpt2-small-chinese-cluecorpussmall`。 我怀疑问题可能并不在于设备的资源上。 ``` (lmflow) [root@a4113ca43b08 LMFlow-main]# ./scripts/run_finetune.sh [2023-04-15 16:04:25,114] [WARNING] [runner.py:186:fetch_hostfile] Unable to find hostfile, will proceed with training with local resources only. [2023-04-15 16:04:25,127] [INFO] [runner.py:550:main] cmd...
Can I customize the indicators related to my data set in this file? #!/bin/bash CUDA_VISIBLE_DEVICES=0 \ deepspeed examples/evaluate.py \ --answer_type medmcqa \ --model_name_or_path /root/autodl-tmp/model/BEELE/ \ --dataset_path /root/autodl-tmp/LMFlow/data/alpaca/test \ --deepspeed examples/ds_config.json...
Why is the lora_model so large after fine-tuning? 
I have question about deepspeed. Assuming that we did not use DeepSpeed in our code initially, if we now want to use it in our fine-tuning process, do we just...
When I read the parameters to be set to finetune.py, I am a little confused. since there are several parameters about evaluation during training: -- validation_file: I did not find...