llm-foundry icon indicating copy to clipboard operation
llm-foundry copied to clipboard

LLM training code for Databricks foundation models

Results 267 llm-foundry issues
Sort by recently updated
recently updated
newest added

When I am doing finetune llama3.1 does the following error occurs, can't locate the exact error, how to fix it please? Running environment: ``` Python 3.11.0rc1 GPU: 2xA100 CUDA Version:...

question

Most HF models have `use_cache` set to `True` by default, which is manually changed to `False` in llm-foundry (most likely due to https://github.com/huggingface/transformers/issues/28056). This PR sets `use_cache` back to True...

Adding temperature tuning in attention similar to https://github.com/huggingface/transformers/blob/9a4ce6477019358abc3ebd72d435da56f4c0ab7c/src/transformers/models/llama4/modeling_llama4.py#L332-L337

Hello, I’m running a 7B model with a 32k context size and seeing unexpected memory scaling behaviors. Here’s the situation: - **Config**: same overall setup, only changing `global_batch_size`. - **Case...

## This PR Adds conversion script for pre-tokenized data in a Delta table. ## Testing MCLI IFT and CPT runs trained successfully.

Hello! question: in data_prep if I use --concat_tokens k, its divide into chunks of k tokens my all data, but if I want to just take sample from my data...