stanford_alpaca icon indicating copy to clipboard operation
stanford_alpaca copied to clipboard

Code and documentation to train Stanford's Alpaca models, and generate the data.

Results 224 stanford_alpaca issues
Sort by recently updated
recently updated
newest added

Now that transformers suport LLAMA it might be a good idea to simplify readme/installation instructions for finetuning instead of having to pick out a specific commit from this PR https://github.com/huggingface/transformers/pull/21955

Tested with own fine-tuned 7B alpaca model ``` python inference.py \ --model_name_or_path {model_path} ``` ``` Instruction: Tell me about alpacas. | 2499 | Al | -15.960 | 0.00% | 29886...

Can someone modify code train.py to support CPU? Thank you in advance

I know that in markdown format `### ` is treated as section/subsection title. Are there any relevance there? Edit to make the question clearer: In the training script, the inputs...

I'm worry this code doesn't run when use pre-trained BLOOMZ or mT0 [https://github.com/bigscience-workshop/xmtf]. Have anyone fine-tuned this ?

[E ProcessGroupNCCL.cpp:455] Some NCCL operations have failed or timed out. Due to the asynchronous nature of CUDA kernels, subsequent GPU operations might run on corrupted/incomplete data. [E ProcessGroupNCCL.cpp:460] To avoid...

Hi , I am trying to get commit as mentioned in FINETUNING section , however, the mentioned hash is not present,. fatal: bad object 68d640f7c368bcaaaecfc678f11908ebbd3d6176. is it available in archive...

Hi. I hope you are having a great new year so far. I can't make heads or tails of what I am supposed to do to be able to train...

Someone told me there is a deepspeed training option in the code, can I ask why its not the default? Do we know if it's far faster, and if so,...