llama-recipes icon indicating copy to clipboard operation
llama-recipes copied to clipboard

Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a...

Results 152 llama-recipes issues
Sort by recently updated
recently updated
newest added

# What does this PR do? This PR updates the `llama_guard_version` argument in inference.py as it takes only string Fixes # (issue) ## Feature/Issue validation/testing Please describe the tests that...

### 🚀 The feature, motivation and pitch Looks like Llama3 has capability to call Tools like google/bing search, Would be good to have an example script with prompt template for...

triaged

# What does this PR do? Fixes # (issue) ## Feature/Issue validation/testing Please describe the tests that you ran to verify your changes and relevant result summary. Provide instructions so...

cla signed

# What does this PR do? This PR update some xpu related logic for correct support. Fixes # (issue) ## Feature/Issue validation/testing Please describe the tests that you ran to...

cla signed

### System Info Hello developer, The Llama-3 model was released today. I want to convert this model to a hf model, but when I follow the readme, the following issue...

triaged

### System Info Dockerfile: ![image](https://github.com/meta-llama/llama-recipes/assets/37894838/5597778d-23b0-49ae-9e9a-05563d38a771) ### Information - [ ] The official example scripts - [ ] My own modified scripts ### 🐛 Describe the bug when freezing the top...

triaged

# What does this PR do? Added a feature that allow users to use pytorch profiler or flop_counter to measure the performance during fine-tuning. For pytorch profiler, use --use_profiler to...

cla signed

# What does this PR do? Add a README.md to this directory that links to API providers that support Meta Llama. ## Before submitting - [X] This PR fixes a...

cla signed

### System Info pytorch2.0.1 cuda11.8 gpu 3090 ### Information - [ ] The official example scripts - [x] My own modified scripts ### 🐛 Describe the bug here's my hyperparameters...

When fine-tuning the 70b model, I always run into an error while loading the model. Usually, after loading 4 to 10 shards (totally15 shards), the following error occurs(see Error Message)....

triaged