FastChat icon indicating copy to clipboard operation
FastChat copied to clipboard

How to finetune mistral-7b with fastchat?

Open Jerry-hyl opened this issue 1 year ago • 1 comments

When I use fastchat to finetune llama2, everything is ok. But when I want to finetune mistral, it shows that "transformer layer not found". I know the main reason is that the value of --fsdp_transformer_layer_cls_to_wrap should not be LlamaDecoderLayer. But what should it be set? Hope you can solve my problem. Many thanks!

Jerry-hyl avatar Jul 22 '24 10:07 Jerry-hyl

same question, please! I tried to change the LlamaDecoderLayer to MistralDecoderLayer, but there are too many warnings or performance drops quickly after sft. Hope for solution.

zouyingcao avatar Jan 17 '25 15:01 zouyingcao