LMFlow icon indicating copy to clipboard operation
LMFlow copied to clipboard

An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.

Results 177 LMFlow issues
Sort by recently updated
recently updated
newest added

Hello, if I want to train vicuna for multiple rounds of dialogue, how should I format the data set? Can you give me an example? Thanks

I was wondering why the gpt2-large model I downloaded from huggingface was 3.1G, but after run_fineturn_with_lora_save_aggregated_weightss.sh it was only 1.5G. This may be a defect in my professional knowledge. Sorry...

pending

Run /scripts/run_raft_align.sh in docker and get an error. deepspeed.runtime.zero.utils.ZeRORuntimeException: You are using ZeRO-Offload with a client provided optimizer () which in most cases will yield poor performance. Please either use...

pending

After I finished fine-tuning with the full parameters, instead of using your chatbot, I called it in the way of pipeline. Why would I output my input in answer every...

pending

"We tested the speculative inference using the first 100 inputs from alpaca test dataset as prompts. When model=gpt2-xl, draft_model=gpt2". I want to test speedup for my own model and draft_model....

Hi, is there any example of python code that can run inference instead of using the command line? If so can you kindly share

Is there a way we can use the trained model using lmflow to integrate Ollama to perform inference

I tried to finetune model with the following scripts: ``` bash scripts/run_finetune.sh ``` As I can see, the commands in this script was like: ``` #!/bin/bash # Please run this...

感觉是baicihuan2的tokenizer做了更新,可能需要适配一下 报了个这样的错误: Tokenizer class BaichuanTokenizer does not exist or is not currently imported. 版本的话是用的最新的代码,v0.05。 感觉改一下tokenizer应该就可以。

pending

When fine-tuning Mistral with LoRA, do you think FlashAttention2 helps in speeding up the process? If yes, how significant is the acceleration? Where is the primary acceleration achieved?