stanford_alpaca
stanford_alpaca copied to clipboard
forward prefetch error: transformers version
Hi,
thanks for your great work! which transformers did you use for training? I tried to reproduce the result with transformers 4.28.0, but I got the following error:
__init__() got an unexpected keyword argument 'forward_prefetch' in transformers/trainer.py. Thank you for your help.
same error
same error
TypeError: init() got an unexpected keyword argument 'forward_prefetch'
@sallywang147 @447428054 Are you fix the error?
May be torch version is not compatible with transformers
https://pytorch.org/docs/1.12/fsdp.html
pytorch version <= 1.12, The FullyShardedDataParallel not have forward_prefetch.
@Jacksonlark @sallywang147 i fix it by installing transformers==4.26.1, I'll post the code in a moment
@Jacksonlark @sallywang147 i fix it by installing transformers==4.26.1, I'll post the code in a moment
@447428054 Using transformers==4.26.1, How do you load llama model?
@447428054 yeah same question as @Jacksonlark. I can't load the model without the latest transformer
@447428054 yeah same question as @Jacksonlark. I can't load the model without the latest transformer
You can check new branch from transformers==4.26.1 and migration llama code, then pip this branch.
Ah thanks! Actually switching torch version worked for me. I still used the latest transformer 4.28.0dev0.
@sallywang147 How are you getting it to fine-tune? My GPU's seem to freeze up after a few minutes, I see no communication or usage, just memory usage and my entire pipeline seems stuck
@447428054 yeah same question as @Jacksonlark. I can't load the model without the latest transformer
You can check new branch from transformers==4.26.1 and migration llama code, then pip this branch.
Yeah, The cod is as follows:https://github.com/447428054/stanford_alpaca_llama.git
Ah thanks! Actually switching torch version worked for me. I still used the latest transformer 4.28.0dev0.
what version of torch is used by you?
Same question. What version of torch?