LMFlow
LMFlow copied to clipboard
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
Make lots of packages optional
**To Reproduce** bash scripts/run_finetune_with_qlora_9_24_4.sh #!/bin/bash # Please run this script under ${project_id} in project directory of # Parses arguments #model_name_or_path=LLM-Research/Meta-Llama-3-70B-Instruct model_name_or_path=/data/hf_cache/hub/models--meta-llama--Meta-Llama-3-70B-Instruct/snapshots/5fcb2901844dde3111159f24205b71c25900ffbd #lora_model_path=/data/midreal/rio/lora/opening_0923 dataset_path=/data/midreal/rio/LMFlow-main/data/opening conversation_template=llama3 output_dir=/data/midreal/rio/LMFlow-main/output_models/qlora_finetuned_llama3_70b_opening_model_9_27 deepspeed_args="--master_port=13001 --include localhost:1" # Safety...
# Description LISA now supports multi GPU. Key points: 1. When initialize models, ds will wrap all model params with optimizer, which uses huge amount of gpu mem. However, according...
I have known that LISA's core code in src\lmflow\pipeline\finetuner.py, mainly in class DynamicLayerActivationCallback. I read it with Algorithm 1 Layerwise Importance Sampling AdamW (LISA) in paper aside. So where is...
I tried to pass --optim , nothing happens. How can I use optimizers such as adamw_8bit or adafactor in LISA? They are not in the custom_optimizers either.
Can lisa training support MiniCPM-V?
Excuse me, I want to know what layers Lisa is frozen in and his freezing strategy. Which file code should I look at?
Does it support qwen2-1.5b?
Is there any Lisa that supports llama3.1 yet?