qlora icon indicating copy to clipboard operation
qlora copied to clipboard

Trained model output seems illegible

Open KKcorps opened this issue 2 years ago • 3 comments

I trained both redpajama-3b and llama-7b for about 1000 iterations.

The loss converged to a pretty low value (0.3) by this time.

But when I load the adapter and try to predict something (using inference notebook code as well as generate.py), I get random unicode tokens in my ouput. Then english answers are not related too much to the original code.

What can be the reason for this?

KKcorps avatar May 28 '23 14:05 KKcorps

7B is a tiny model. You cannot expect any quality in the output.

phalexo avatar May 28 '23 14:05 phalexo

No I meant it should still be slightly relatable and not filled with gibberish tokens.

KKcorps avatar May 28 '23 14:05 KKcorps

@KKcorps were you able to find the issue and get any meaningful output from the trained models?

amdnsr avatar May 28 '23 21:05 amdnsr

@KKcorps this happened when my learning rate was too high.

Birch-san avatar Jul 03 '23 10:07 Birch-san

@KKcorps this happened when my learning rate was too high.

I was able to solve this using a different base model. but if learning rate it too high then the training/validation loss should also not keep on dropping right?

KKcorps avatar Jul 03 '23 10:07 KKcorps