Shoaib Hossain
Shoaib Hossain
I've figured out the **solution**. Below is the code for those who just want the solution, not the details: ### Solution Code: ``` # change the padding tokenizer value tokenizer.add_special_tokens({"pad_token":...
> @danielhanchen in colab after finetuning I was having the same issue and created an issue #416 . I've posted a solution [here](https://github.com/unslothai/unsloth/issues/416#issuecomment-2094745798).
> @KillerShoaib Man thanks a lot for fixing, I really appreciate that. Can u explain to me where to add that I am using Google Colab t4 GPU: https://colab.research.google.com/drive/135ced7oHytdxu3N2DNe1Z0kqjyYIkDXp?usp=sharing as...
> @danielhanchen @KillerShoaib I checked it once again it literally the same  Okay, I think you're using your finetuned model which was finetuned on top of old unsloth...
> @KillerShoaib Iam using this colab notebook : https://colab.research.google.com/drive/135ced7oHytdxu3N2DNe1Z0kqjyYIkDXp?usp=sharing from their github I am using llama3-8b-instruct model : https://huggingface.co/unsloth/llama-3-8b-Instruct I've just downloaded **unsloth/llama-3-8b-Instruct** and verified its pad token and eos...
> @KillerShoaib Its the same stuff dude!! just generates sometimes and goes on loop like this  till 128 max new tokens > > and if text streaming is true...