llama-recipes icon indicating copy to clipboard operation
llama-recipes copied to clipboard

Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a...

Results 272 llama-recipes issues
Sort by recently updated
recently updated
newest added

I hope this message finds you well. I recently had the opportunity to experiment with the Codellama-7b-Instruct model from GitHub repository and was pleased to observe its promising performance. Encouraged...

triaged

### System Info PyTorch version: 2.0.1+cu117 Is debug build: False CUDA used to build PyTorch: 11.7 ROCM used to build PyTorch: N/A OS: Red Hat Enterprise Linux release 8.8 (Ootpa)...

bug

Sorry for the stupid question, I'm not sure why I can't seem to investigate this well, and it's driving me nuts. I've tried running the example finetuning script on the...

triaged

### 🚀 The feature, motivation and pitch Hi all! Has anyone edited the code so to support logging on Tensorboard or WanDB? Thanks! ### Alternatives _No response_ ### Additional context...

### System Info 2.0.1+cu118, driver 535.86.05 ### Information - [X] The official example scripts - [ ] My own modified scripts ### 🐛 Describe the bug When you run on...

### 🚀 The feature, motivation and pitch i follow the instructions like python llama_finetuning.py \ --use_peft \ --quantization \ --model_name "meta-llama/Llama-2-7b-chat-hf" \ --output_dir Path/to/save/PEFT/model i wander if i do not...

### System Info ``` torch=2.0.1+cu118 NVIDIA TITAN RTX 3090 NVIDIA-SMI 525.116.04 Driver Version: 525.116.04 CUDA Version: 12.0 ``` ### Information - [X] The official example scripts - [ ] My...

question
triaged

### System Info colab GPU:v100 ### Information - [ ] The official example scripts - [ ] My own modified scripts ### 🐛 Describe the bug ![微信截图_20230829170647](https://github.com/facebookresearch/llama-recipes/assets/14157458/68c86120-5577-44db-81ba-af26ced834a7) ### Error logs...

triaged

I want to continue pre-training the LLMa70b model in order to add Chinese tokens and train it on Chinese data. I'm considering using FSDP along with pure bf16. However, I...

question

Hi, I am going to do distributed training of llama on aws sagemaker as managed training across multiple devices/nodes. Sagemaker provides data parallel and model parallel distributed training in sagemaker....