xDAN-AI
xDAN-AI
`Fetching 31 files: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 31/31 [00:00
### Describe the issue Issue: Command: ``` Bash pretrain.sh on my fineunted Llama2 model. ``` Log: ``` You should probably TRAIN this model on a down-stream task to be able...
### Please check that this issue hasn't been reported before. - [X] I searched previous [Bug Reports](https://github.com/OpenAccess-AI-Collective/axolotl/labels/bug) didn't find any similar reports. ### Expected Behavior ### Training in Mixtral model...
is it toooo long to quantized a model ?
` During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/workspace/LLaMA-Factory/src/train_bash.py", line 14, in main() File "/workspace/LLaMA-Factory/src/train_bash.py", line 5, in main run_exp() File "/workspace/LLaMA-Factory/src/llmtuner/train/tuner.py",...
training env: LLaMaFactory `01/24/2024 01:53:50 - INFO - llmtuner.model.patcher - Quantizing model to 4 bit. Traceback (most recent call last): File "/usr/local/lib/python3.10/dist-packages/transformers/utils/versions.py", line 102, in require_version got_ver = importlib.metadata.version(pkg) File...
raise ValueError(f"{tensor_name} is on the meta device, we need a `value` to put in on {device}.")
checkpoint shards: 73%|███████▎ | 11/15 [00:02
I found the sglang worker inside, supported?
### Reminder - [X] I have read the README and searched the existing issues. ### Reproduction 1. After merging the models, I found that two parameters in token.config were not...
vLLM had supported awq quantized model yet. Please add one more params to set --quantization awq