ibicdev
ibicdev
Like this idea. For now, can you display semantic kernel's plan using chainlit's TaskList? https://docs.chainlit.io/api-reference/elements/tasklist
Any update on this? Does the PR above add this feature?
I am also using cuda 11.8, and pytorch 2.01 for cuda 11.8. Also tried pytorch nightly and got the same error. `--use_flash_attn False` didn't make a difference either. The error...
Yes, it worked well without any code change.
The only change I made is `--model_id`, from `tiiuae/falcon-180B` to `meta-llama/Llama-2-70b-hf`. The full command is ``` torchrun --nproc_per_node 8 run_ds_lora.py \ --model_id meta-llama/Llama-2-70b-hf \ --dataset_path dolly-processed \ --output_dir falcon-180b-lora-fa \...
Ahh, I didn't. I saw your code https://github.com/philschmid/deep-learning-pytorch-huggingface/blob/main/training/utils/peft_utils.py#L38-L41, and thought it's already taken care of. Also, even when I used `--use_flash_attn False` I still got the same error.
Excited to see flash-attn 2 natively supported in transformers! Would you plan to update this post to work with this new feature?
Great! Looking forward to the updates.