silvestron

Results 11 comments of silvestron

Is it okay to use llama for this model? When I tried to set model type to mpt I got his error: ``` ERROR:Unknown pre-quantized model type specified. Only 'llama',...

@jpturcotte Can you run `git show` inside the text-generation-webui folder to see what commit you're on? How do you get to run it though? I have to manually specify the...

@NickWithBotronics `"use_cache": false` in the config file, I haven't touched it. How much does the webui respect the config file though? It doesn't seem to care about the model type,...

I replaced `git pull` with `git checkout 85238de` in webui.py, but it looks like going back to older commit breaks things. Maybe running a clean install with that commit would...

I'd also add that on a working, up to date installation, I tried using llama, gptj, and opt as model type and gave the same results.

Are we talking about the 4bit model? That doesn't work if you don't specify a model. #1894 I get the same error if I don't give it a model type.

That actually made the token generation faster, however the initialization time, that on my hardware takes 9 minutes didn't change. The config has `"init_device": "cpu"` and the console says you...

@jpturcotte All good, I couldn't do much with this model anyway without much of VRAM anyway. I guess multi-GPU is going to be the only way to run models that...

I guess I can reinstall, but I wonder what broke my current installation.

I did a clean install but that didn't solve the problem.