The_Stallion
The_Stallion
I tried with a different model and it works fine `from peft import LoraConfig, TaskType peft_config = LoraConfig(task_type=TaskType.SEQ_2_SEQ_LM, inference_mode=False, r=8, lora_alpha=32, lora_dropout=0.1) from transformers import AutoModelForSeq2SeqLM model = AutoModelForSeq2SeqLM.from_pretrained("bigscience/mt0-large") from...
I am using qlora with 4-bit quant but somehow i have the same error. For more detail, this is the config i used : `BitsAndBytesConfig { "_load_in_4bit": true, "_load_in_8bit": false,...
Hello, `pip install networkx-metis` and https://pypi.org/project/networkx-metis/ doesn't work for me neither. Plus i tried to clone the repo and build -> install and i got compilation errors with Cython. (I...
> Hi @dipanjanS ! Thanks for the issue, I had a deeper look. Previously there was a silent bug in transformers that was quantizing the `pre_classifier` layer, which shouldn't happen...