Yili Hong
Yili Hong
I have already applied for access to meta llama3. And I also receive the email about how can I use download scripts to download it, but I still cannot have...
> In our model, we share the same weight matrix between the two embedding layers and the pre-softmax linear transformation. Hello! I have the paper recently and find that this...
I want to test the ability of llama2-7b-hf instead of llama2-7b-chat. But llama2-7b-hf is a base model not having a chat template, so how can I use llama2-7b-hf and transformers...
When I use fastchat to finetune llama2, everything is ok. But when I want to finetune mistral, it shows that "transformer layer not found". I know the main reason is...
When I use training code to finetune mistral-7b,I have met the exception like "Could not find the transformer layer class to wrap in the model.". How can I fix it?