alpaca-lora icon indicating copy to clipboard operation
alpaca-lora copied to clipboard

training results of alpaca-lora 7b and original llama 7b

Open frankzhao112 opened this issue 1 year ago • 1 comments

image 1 first we load both llama 7b and aplaca-lora-7b . image 2 u can use colab to test both llama -7b and aplaca-lora-7b , u can not find any improvements at all. Using int8 type and peft lora training llama significantly reduces the cost of gpu memory , but the training results can not satisfy the needs of a online chat system.

3 after I use other datasets to train llama using int8 type and peft lora traning, I still get the same results.

frankzhao112 avatar Apr 28 '23 02:04 frankzhao112

I tried to fine-tune with my own dataset. Similarly, it seems not learning much. Should I adjust the learning rate or parameters for LORA? Any advice?

BinWang28 avatar May 05 '23 04:05 BinWang28