DeepSeek-Coder icon indicating copy to clipboard operation
DeepSeek-Coder copied to clipboard

Catastrophic forgetting problem

Open shatealaboxiaowang opened this issue 1 year ago • 2 comments

Hi dear:

Thanks for your open source, but when i finetuned (whatever full parameters or LoRa ) on my dataset, catastrophic forgetting kept coming up (decrease in performance on the humaneval), i do not know how to solve it, do you have any tops?

shatealaboxiaowang avatar Mar 06 '24 09:03 shatealaboxiaowang

If you fine-tune on the Instruct model, I think such a phenomenon is normal. There are two reasons for this: one is that your data is inferior compared to the original data used to fine-tune the Instruct model, and the other reason is that further fine-tuning on the Instruct model could lead to overfitting.

guoday avatar Mar 12 '24 02:03 guoday

Thx, i will continue to check my data quality.

shatealaboxiaowang avatar Mar 19 '24 06:03 shatealaboxiaowang