stanford_alpaca icon indicating copy to clipboard operation
stanford_alpaca copied to clipboard

Code and documentation to train Stanford's Alpaca models, and generate the data.

Results 224 stanford_alpaca issues
Sort by recently updated
recently updated
newest added

Thanks for the great work, I reproduced the training, but at inference time tends to generate shorter text. I am using: `generated = model.generate(batch["input_ids"], max_length=512)` Does the interface on the...

Added explicit instructions. ``` pip uninstall transformers pip install git+https://github.com/zphang/transformers.git@68d640f7c368bcaaaecfc678f11908ebbd3d6176 ```

Hello, first of all thank you for releasing the training code for alpaca, we really appreaciate it. I am running the fine-tuning script on an 4xA100-SXM4-80GB, and currently getting an...

In the provided training command: ```bash torchrun --nproc_per_node=4 --master_port= train.py \ --model_name_or_path \ --data_path ./alpaca_data.json \ --bf16 True \ --output_dir \ --num_train_epochs 3 \ --per_device_train_batch_size 4 \ --per_device_eval_batch_size 4 \...

Once you collect 52k synthetic dataset, how did you plot the pie chart [here](https://github.com/tatsu-lab/stanford_alpaca/blob/main/assets/parse_analysis.png) ? Thanks !

Hi, is there any chance there will be a version of the data generator that supports the gpt-3.5-turbo model?

Can this finetuning script fit into A10, which only has 24GB GPU memory? I am trying to fine-tune the model on 4 A10 GPUs using a batch size of 1,...

My first run of the trainer could not save the model because the evaluate() call fails. I have removed that method call and now would like to resume from the...