stanford_alpaca icon indicating copy to clipboard operation
stanford_alpaca copied to clipboard

Code and documentation to train Stanford's Alpaca models, and generate the data.

Results 228 stanford_alpaca issues
Sort by recently updated
recently updated
newest added

I am working on a project using this model, but I hit a roadblock by not knowing how to provide it with extra context as a pdf or text file....

Hi, i try to run train.py in Windows. Help me please solve the problem. # System parameters 12th Gen Intel(R) Core(TM) i5-12600KF 3.70 GHz 32 GB Cuda 11.8 Windows 11...

Has anyone been able to finetune any of the models larger than 7b successfully? I'm training on 8 A100s with 80GB of RAM each which is more than enough space....

I don't understand why the default model_max_length is set to 512, and the example training bash script on the main readme doesn't pass in an argument for that as 2048...

I may be wrong here but from what I traced from their train.py file, it seems like they are training the model by passing both question + answer as the...

Thanks for the repo. If I use torchrun as suggested by the webpage (see below) it fails due to an error while compiling cpuadam within the deepspeed library. The actual...

A month ago, I train the alpaca with 4 A100 GPUs (each 80G) and `per_device_train_batch_size=4`. Here `transformers==4.28.1`. Today I retrain the alpaca with the same hardwares and the same code,...

I have been trying to finetune for a while and I succeeded using this repository. However, the next day I tried to finetune again using the same steps and it...

Can I fine tuning the alpaca on the machine with 4 * V100 (32G)?

I cannot find any code to define the wandb to visulize the training process in train.py. Can anyone answer my question?