alpaca-lora
alpaca-lora copied to clipboard
Differences between this code and zphang's minimal-llama trainig code
Originally I was playing around with https://github.com/zphang/minimal-llama/ to generate alpaca-like adaptation here. Both use peft but in a slightly different fashion, like different parameter saving and using a custom trainer. Was wondering which one is the idiomatic approach?