localGPT
                                
                                 localGPT copied to clipboard
                                
                                    localGPT copied to clipboard
                            
                            
                            
                        How to load a quantized model ?
How do I load the following model? TheBloke/vicuna-7B-1.1-GPTQ-4bit-128g
When I try I get he following error: TheBloke/vicuna-7B-1.1-GPTQ-4bit-128g does not appear to have a file named pytorch_model.bin, tf_model.h5, model.ckpt or flax_model.msgpack.
How Do I pass the correct parameters for the model to work?
As far as I know, quantized models are not yet supported. I might be wrong.
Yes, it looks like they are not supported, yet.
This is mentioned here: https://github.com/PromtEngineer/localGPT/issues/111#issuecomment-1583131451