LLaVA icon indicating copy to clipboard operation
LLaVA copied to clipboard

[Question] pretain 7B loss doesn't change

Open aprilehannibal opened this issue 1 year ago • 3 comments

Question

I tried to pretrain llava-7b on 8 * A100(40G), CUDA117, torch==1.13.1, py3.10 I found the loss doesn't change when training.

image

here is my pretraining scripts: image

aprilehannibal avatar May 17 '23 15:05 aprilehannibal

I wonder the LLM you used for pretraining is the "lmsys/vicuna-7b-delta-v1.1" or the origin "llama 7b" weight. @haotian-liu

aprilehannibal avatar May 17 '23 15:05 aprilehannibal

Hi @aprilehannibal, you need to perform weight conversion for vicuna weights, by combing llama-7b and vicuna-delta -> vicuna-7b. We pretrain based on vicuna-7b.

We'll update the instruction to make it clearer, thanks.

haotian-liu avatar May 17 '23 16:05 haotian-liu

OK, got it! Thanks a lot!

aprilehannibal avatar May 18 '23 02:05 aprilehannibal