Ryusaeba
Ryusaeba
@haotian-liu This is similar issue as what FastChat meet. The root cause is Huggingface introduce some bugs when dealing with added tokens. Please refer the fix [here](https://github.com/lm-sys/FastChat/pull/2498).
@kperi Did you happen to have the 1000 LIMA training data?
Are you using 0.1 of weight for the data with unknown correctness and 1.0 of weight for correct one? If not, could you please reveal more details?
Understood. Please share with me if you have any update. Also looking forward your expansion to Mistral and Multi-Modal models.
@SunMarc, Will be there has a patch release for v4.45 series?
Thank you @SunMarc. I tried with v4.45.2 and the issue still persist. Will give it a try with latest transfromers.
The issue was happened at Gemma-2. I will see whether we can prepare a scripts for reproducible.
@SunMarc The issue is still persist. Please see the following code and help on this issue. **CODE** ``` from transformers import ( AutoModelForCausalLM, AutoTokenizer, ) import pdb MODEL_PATH='/llm_data2/huggingface/models/google/git_version/gemma-2-2b-it' texts =...
@SunMarc Understood. Thanks for the workaround. Will give it a try. We are using [run_clm.py](https://github.com/huggingface/transformers/blob/main/examples/pytorch/language-modeling/run_clm.py) for finetuning experiments, it would be great if the transformers library can integrate the fix....