peft
peft copied to clipboard
Fine-tuning a 13B mt0-xxl model
Hello and thanks for the awesome library !
I'd like to reproduce some of the results you display in the repo's README and had a few questions:
- I was wondering what you meant by
Also, we didn't use the larger 13B mt0-xxl model
as written in the Use cases
section of the README ?
- What code did you use for the
PEFT-LoRA PyTorch
xmt0-xxl
datapoint? I'm assuming it's the one fromexamples/conditional_generation/peft_lora_seq2seq.ipynb
but I run into OOM errors when I try on a similar setup (i.e. A100 80GB + fair amount of RAM).
Thanks in advance :)