bartman081523

Results 38 comments of bartman081523

> This is the output I get from conda list -p "C:\Users\patri\miniconda3\envs\textgen": Your env looks alright, as far as i can tell. > Wait - there is something wrong here....

I found this maybe relevant: https://github.com/huggingface/peft/blob/main/examples/causal_language_modeling/peft_lora_clm_accelerate_big_model_inference.ipynb ``` from peft import PeftModel, PeftConfig max_memory = {0: "1GIB", 1: "1GIB", 2: "2GIB", 3: "10GIB", "cpu": "30GB"} peft_model_id = "smangrul/twitter_complaints_bigscience_bloomz-7b1_LORA_CAUSAL_LM" config = PeftConfig.from_pretrained(peft_model_id)...

I tried without the "--gptq-bits 4", that failed with another error: ``` python server.py --model llama-7b --lora alpaca --listen --gpu-memory 11 --cpu-memory 16 --disk ===================================BUG REPORT=================================== Welcome to bitsandbytes. For...

@wywywywy @BadisG found a way to fix 4-bit mode: https://github.com/oobabooga/text-generation-webui/issues/332#issuecomment-1474883977 change the `lora.py` from the `peft` package: `C:\Users\Utilisateur\anaconda3\envs\textgen\lib\site-packages\peft\tuners\lora.py` in Linux `venv/lib/python3.10/site-packages/peft/tuners/lora.py` fixed `lora.py` https://pastebin.com/eUWZsirk @BadisG added those 2 instructions on...