LLaMA-Pro icon indicating copy to clipboard operation
LLaMA-Pro copied to clipboard

Training on arbitary data

Open HelloWorldLTY opened this issue 1 year ago • 2 comments

Thanks for your great work. I wonder if it is possible to directly use alpaca_lora or stanford_alpaca to finetune 8B model in arbitary dataset. Can we access the code? Or we directly use block_expand to create a new model and then train that new model? Does this support huggingface version? Thanks.

HelloWorldLTY avatar Apr 08 '24 13:04 HelloWorldLTY

Yes, you can directly finetune the 8B model with any datasets. You can access the model in the huggingface (https://huggingface.co/TencentARC/LLaMA-Pro-8B). You can use it just like the normal LLaMA model.

hills-code avatar Apr 08 '24 13:04 hills-code

Thanks, will try.

HelloWorldLTY avatar Apr 08 '24 15:04 HelloWorldLTY