stanford_alpaca
stanford_alpaca copied to clipboard
Generation problem after / before instruction fine-tuning
Environment: 6xA6000 48GB with Ubuntu 22.04, Pytorch 1.13.0
I ran into a generation problem after following your instruction to convert LLaMA-7B weight using your attached script.
I simply used the following script to directly test generation after loading the converted LLaMA-7B model:
tokenizer.batch_decode(model.generate(**tokenizer('I want to ', return_tensors="pt")))
The output of above code is:
'I want to acoérницschutzirectorioieckťDEX threshold släktetolasĭüttpiel'
The problem happens both before and after following your README for instruction fine-tuning. (note that I see the loss is decreasing over time during the fine-tuning stage which seems OK)
I have no problem running generation using original code from LLaMA, may I know your generation script so that I can test what caused the problem? Thanks.