devzzzero
devzzzero
Yep. That did it. Thank you.
> > @davedgd Oh so Unsloth is fine (the models or just finetuning with Unsloth?) but the Meta ones still don't work as expected? > > Correct, but to clarify,...
> See https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/commit/4d6c61da057c45bfc4dc4d3bfa5a691ecb9ce0cf > > Yes the pad token is in fact a bug fix Indeed. My pull of the official Llama3 hf models occurred more than 20 days ago...
> There are two problems in your code. First, the llama-3 chat template itself introduces eos_token at the end of every system/user/assistant prompt, so initialize **pad_token = eos_token** is a...
Its running now, ` per_device_train_batch_size = 1` :-( ETA ~15 hours