llm-foundry icon indicating copy to clipboard operation
llm-foundry copied to clipboard

LLM training code for Databricks foundation models

Results 267 llm-foundry issues
Sort by recently updated
recently updated
newest added

## 🚀 Feature Request Allow passing `num_proc`/`num_workers` parameter in`InContextLearningDataset` so that preparation of dataset can use more than one processes. ## Motivation When loading bigger ICL eval datasets, it is...

enhancement

could run the llm-foundry on AMD 4xMi250 machine Steps to reproduce the behavior: 1. follow latest instructions from: https://github.com/ROCm/flash-attention/tree/flash_attention_for_rocm start from docker image: rocm/pytorch:rocm5.7_ubuntu22.04_py3.10_pytorch_2.0.1 export GPU_ARCHS="gfx90a" export PYTHON_SITE_PACKAGES=$(python -c 'import...

bug

Previously the dataloader will throw the error message that contain duplicated fileName + extension in the `Could not find a file with any of the supported extensions:`. This PR patched...

I was wondering if there was a straightforward way to convert from sharded to monolithic checkpoint for a subsequent conversion to hf format (not a direct conversion sharded -> hf)....

question

In order to support customers with sensitive storage network configurations, we have to use the `log_model` API. This will cause duplicate artifact uploads, which is not efficient, so we will...

At TensorWave we're actively using LLM-Foundry on our MI300X systems with great success, since ROCm version 6.1. What information do you need to see in order to add MI300X to...

bug

Updates the requirements on [transformers](https://github.com/huggingface/transformers) to permit the latest version. Release notes Sourced from transformers's releases. Patch Release v4.45.1 Patches for v4.45.1 [MllamaProcessor] Update errors and API with multiple image...

dependencies