llama-recipes
llama-recipes copied to clipboard
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a...
## Main Goal Building an e2e recipe for building chatbots. ### High level idea We want to focus on following stages: - Data pipelines for creating datasets for chatbots -...
### System Info [pip3] numpy==1.26.3 [pip3] torch==2.2.0+cu118 [pip3] triton==2.2.0 [conda] numpy 1.26.3 pypi_0 pypi [conda] torch 2.2.0+cu118 pypi_0 pypi [conda] triton 2.2.0 pypi_0 pypi GCP instance g2-standard-48, 4xL4 GPU ###...
### System Info python eval.py --model hf --model_args pretrained=/mnt/sdb/zjk/llama2/llama-recipes/Llama-2-7b-hf,dtype="float",peft=/mnt/sdb/zjk/llama2/llama2-lora --tasks hellaswag --num_fewshot 10 --device cuda:0 --batch_size 8 error: 2024-04-02:06:58:12,156 ERROR [eval.py:226] An error occurred during evaluation: module 'lm_eval.tasks' has no...
### System Info 路径错误 ### Information - [X] The official example scripts - [ ] My own modified scripts ### 🐛 Describe the bug from lm_eval.utils import make_table改为from lm_eval.evaluator import...
# What does this PR do? - Adds a note on using the correct CUDA version for pip installs. - Removes 'test' from the pytorch whl install
# What does this PR do? Remove openai from example notebook and llm.py class. Simplify notebook layout and highlight additional packages that need to be installed. As discussed with @HamidShojanazeri...
This is add the implementation of H2O algorithm for efficient long context inference of Llama models. Current implementations are based on the Huggingface transformers and tests on summarization tasks, including...
BUG2
### 🚀 The feature, motivation and pitch I propose adding a feature to improve the visualization of loss curves in LLaMA-Recipes. Currently, the loss curves may exhibit spikes or irregularities,...
### System Info PyTorch version:2.2.1 ### Information - [X] The official example scripts - [X] My own modified scripts ### 🐛 Describe the bug The pretraining loss curve in LLaMA2...
Dear LLaMA Teams, A huge thank you for making your remarkable work available to the public! I've taken a close look at the pretraining loss curves depicted in Figure 1...