llama-recipes icon indicating copy to clipboard operation
llama-recipes copied to clipboard

Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a...

Results 272 llama-recipes issues
Sort by recently updated
recently updated
newest added

### 🚀 The feature, motivation and pitch I am running 3090 with 24GB VRAM and 16GB Shared memory ( Total is 40 GB ) When i am fine tuning 7B...

I am wondering how we could adapt the example.py files provided to undertake tasks such as: 1. Identify the extent of positive sentiment and negative sentiment in the following text....

### System Info Cuda 12.1 PyTorch 2.3.0 Python 3.11 ``` Thu May 23 15:30:20 2024 +---------------------------------------------------------------------------------------+ | NVIDIA-SMI 545.23.06 Driver Version: 545.23.06 CUDA Version: 12.3 | |-----------------------------------------+----------------------+----------------------+ | GPU Name...

Hi, I was using llama recipe [local inference](https://github.com/meta-llama/llama-recipes/tree/main/recipes/inference/local_inference). But I get the warning: `Token indices sequence length is longer than the specified maximum sequence length for this model (1998 >...

### System Info [pip3] numpy==1.26.3 [pip3] torch==2.3.1+cu121 [pip3] torchaudio==2.3.1+cu121 [pip3] torchvision==0.18.1+cu121 [pip3] triton==2.3.1 [conda] numpy 1.26.3 pypi_0 pypi [conda] torch 2.3.1+cu121 pypi_0 pypi [conda] torchaudio 2.3.1+cu121 pypi_0 pypi [conda] torchvision...

# What does this PR do? Updated the endpoint to 3.1 support. Also updated Langchain and Gradio support as their framework updated. ## Before submitting - [X] This PR fixes...

cla signed

### System Info This is independent of torch version. ### Information - [X] The official example scripts - [ ] My own modified scripts ### 🐛 Describe the bug The...

# What does this PR do? This PR refactors the existing script to improve modularity, readability, and error handling. [src/llama_recipes/inference/checkpoint_converter_fsdp_hf.py] ### Description of the Change - **Modularization**: Introduced `get_model_name_from_yaml` and...

### System Info PyTorch version: 2.2.2+cu121 Is debug build: False CUDA used to build PyTorch: 12.1 ROCM used to build PyTorch: N/A OS: Debian GNU/Linux 10 (buster) (x86_64) GCC version:...

triaged

# What does this PR do? Meta's latest Llama3.1 models offer unique function calling capabilities. In particular they offer built-in tool calling capabilities for the following 3 external tools: *...

cla signed