LMFlow
LMFlow copied to clipboard
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
Hello, if I want to train vicuna for multiple rounds of dialogue, how should I format the data set? Can you give me an example? Thanks
I was wondering why the gpt2-large model I downloaded from huggingface was 3.1G, but after run_fineturn_with_lora_save_aggregated_weightss.sh it was only 1.5G. This may be a defect in my professional knowledge. Sorry...
Run /scripts/run_raft_align.sh in docker and get an error. deepspeed.runtime.zero.utils.ZeRORuntimeException: You are using ZeRO-Offload with a client provided optimizer () which in most cases will yield poor performance. Please either use...
After I finished fine-tuning with the full parameters, instead of using your chatbot, I called it in the way of pipeline. Why would I output my input in answer every...
"We tested the speculative inference using the first 100 inputs from alpaca test dataset as prompts. When model=gpt2-xl, draft_model=gpt2". I want to test speedup for my own model and draft_model....
Hi, is there any example of python code that can run inference instead of using the command line? If so can you kindly share
Is there a way we can use the trained model using lmflow to integrate Ollama to perform inference
I tried to finetune model with the following scripts: ``` bash scripts/run_finetune.sh ``` As I can see, the commands in this script was like: ``` #!/bin/bash # Please run this...
感觉是baicihuan2的tokenizer做了更新,可能需要适配一下 报了个这样的错误: Tokenizer class BaichuanTokenizer does not exist or is not currently imported. 版本的话是用的最新的代码,v0.05。 感觉改一下tokenizer应该就可以。
When fine-tuning Mistral with LoRA, do you think FlashAttention2 helps in speeding up the process? If yes, how significant is the acceleration? Where is the primary acceleration achieved?