stanford_alpaca
stanford_alpaca copied to clipboard
Code and documentation to train Stanford's Alpaca models, and generate the data.
I am working on a project using this model, but I hit a roadblock by not knowing how to provide it with extra context as a pdf or text file....
Hi, i try to run train.py in Windows. Help me please solve the problem. # System parameters 12th Gen Intel(R) Core(TM) i5-12600KF 3.70 GHz 32 GB Cuda 11.8 Windows 11...
Has anyone been able to finetune any of the models larger than 7b successfully? I'm training on 8 A100s with 80GB of RAM each which is more than enough space....
I don't understand why the default model_max_length is set to 512, and the example training bash script on the main readme doesn't pass in an argument for that as 2048...
I may be wrong here but from what I traced from their train.py file, it seems like they are training the model by passing both question + answer as the...
Thanks for the repo. If I use torchrun as suggested by the webpage (see below) it fails due to an error while compiling cpuadam within the deepspeed library. The actual...
A month ago, I train the alpaca with 4 A100 GPUs (each 80G) and `per_device_train_batch_size=4`. Here `transformers==4.28.1`. Today I retrain the alpaca with the same hardwares and the same code,...
I have been trying to finetune for a while and I succeeded using this repository. However, the next day I tried to finetune again using the same steps and it...
Can I fine tuning the alpaca on the machine with 4 * V100 (32G)?
I cannot find any code to define the wandb to visulize the training process in train.py. Can anyone answer my question?