FastChat
FastChat copied to clipboard
Could not find the transformer layer class to wrap in the model.
When I use training code to finetune mistral-7b,I have met the exception like "Could not find the transformer layer class to wrap in the model.". How can I fix it?
Hi Jerry, Did you find a solution for this? I have the same exception
@juliaparedesq I tried to remove --fsdp_transformer_layer_cls_to_wrap 'LlamaDecoderLayer' and the exception was gone. But after finetuning, the model's ability declined significantly. It seems that fastchat can only be used to deploy mistral model instead of finetuning it.