ColossalAI
ColossalAI copied to clipboard
Making large AI models cheaper, faster and more accessible
### π Describe the bug Here are my script, it can run with hybrid_parallel plugin, but other plugins have the same error "out of memory" torchrun --standalone --nproc_per_node 8 finetune.py...
### Describe the feature You are using a model of type mistral to instantiate a model of type llama. This is not supported for all configurations of models and can...
### π Describe the bug RuntimeError: FlashAttention only supports Ampere GPUs or newer. ### Environment colossal 0.3.4 ColossalAI/examples/language/opt# bash run_demo.sh + pip install -r requirements.txt
### π Describe the bug I want to build from source the entire project, and failed on `pip install .`, the error seems to be pytorch header related. I'm using...
## π Checklist before creating the PR - [ ] I have created an issue for this PR for traceability - [ ] The title follows the standard format: `[doc/gemini/tensor/...]:...
## π Checklist before creating the PR - [ ] I have created an issue for this PR for traceability - [ ] The title follows the standard format: `[doc/gemini/tensor/...]:...
[BUG]: How to run llama2 70B pretrain on 32gpus? I got OOM error on almost every plugin and config.
### π Describe the bug gemini / gemini_auto / zero2 / hybrid_parallel I have tried and still got OOM error. with hybrid_parallel plugin , I tried configs as follows: 1....
## π Checklist before creating the PR - [ ] I have created an issue for this PR for traceability - [ ] The title follows the standard format: `[doc/gemini/tensor/...]:...
### Describe the feature To reduce memory usage, support for LoRa is provided.
### Proposal DOC: https://n4fyd3ptax.feishu.cn/docx/MhlmdHsGkoeoslx9fqucPO17n9b ### Self-service - [ ] I'd be willing to do some initial work on this proposal myself.