Marthinus Spangenberg

Results 5 comments of Marthinus Spangenberg

I just stumbled across the same issue - it seems to be changes since in Python 3 removing StringIO and cStringIO modules and replacing them with io.StringIO and io.BytesIO. @vinta...

For reference: https://github.com/hwchase17/langchain/issues/2592#issuecomment-1502065790 The error is caused by the ggml model you're attempting to use not being compatible with the version of llama.cpp being used by the web-ui. If you're...

Use GGML for CPU inference. Try the [WizardLM-7B-uncensored.ggml.q4_0](https://huggingface.co/TheBloke/WizardLM-7B-uncensored-GGML/blob/previous_llama/) model. Just create a folder in your models folder called `WizardLM-7B-uncensored-GGML` and download: `https://huggingface.co/TheBloke/WizardLM-7B-uncensored-GGML/resolve/previous_llama/WizardLM-7B-uncensored.ggml.q4_0.bin` into it. Then start the server with `--cpu...

Did you make sure to download only the model I suggested, the one in the [previous_llama](https://huggingface.co/TheBloke/WizardLM-7B-uncensored-GGML/tree/previous_llama) branch of the repo? The error you're getting suggests you downloaded the files from...

That's great! Glad you could get it working. Here are more models that work with the current version of llama.cpp: - [stable-vicuna-13B-GGML](https://huggingface.co/TheBloke/stable-vicuna-13B-GGML/tree/previous_llama) - [wizard-vicuna-13B-GGML](https://huggingface.co/TheBloke/wizard-vicuna-13B-GGML/tree/previous_llama) - [GPT4All-13B-snoozy-GGML](https://huggingface.co/TheBloke/GPT4All-13B-snoozy-GGML/tree/previous_llama) Also see: https://github.com/oobabooga/text-generation-webui/issues/2020#issuecomment-1546656696