alpaca-lora icon indicating copy to clipboard operation
alpaca-lora copied to clipboard

The output returning input.

Open TemaLykov opened this issue 2 years ago • 1 comments

I finetuned the decapoda and yahma versions of thr alpaca on my custom dataset, but I always got the same result - the generated output looks like: ###Instruction... and so on, repeating my input. I saw issues with the same problem and i see that problem can be with the eos and pad tokens after changing tokenization format. But if i got it right, using last version of tloen/alpaca-lora with: model.config.pad_token_id = tokenizer.pad_token_id = 0 # unk model.config.bos_token_id = 1 model.config.eos_token_id = 2 should have solve the problem. I also used yahma/llama-7b-hf. I saw in one issue that can resolve it. But all my attempts finishes the same way - the repeating of my input. what I misunderstand of doing wrong?

TemaLykov avatar Apr 28 '23 01:04 TemaLykov

i am facing similar issue while finetuning with decapoda although it gives me proper output on some inputs

rogerbarretocode avatar Apr 29 '23 12:04 rogerbarretocode

Hi, you can try my script to test if your model is performing ok: https://github.com/tloen/alpaca-lora/issues/454

I also have inputs in my response when I use the generate.py file, but I did some debugging and it looks like the trained model is performing ok.

DanqingZ avatar May 15 '23 07:05 DanqingZ