Branko Radovanović
Branko Radovanović
### System Info GPT4All 2.4.4 AVX Windows 10 x64 ### Information - [ ] The official example notebooks/scripts - [ ] My own modified scripts ### Related Components - [...
It seems that the message "Recalculating context" in the chat (or "LLaMA: reached the end of the context window so resizing" during API calls) appears after 2k tokens, regardless of...
### Feature request In the chat app, there is no way (that I've found) to select and/or copy an entire _response_. One can: 1. Select a single _word_ (by double-clicking)...
### System Info GPT4All 2.5.2 ### Information - [ ] The official example notebooks/scripts - [ ] My own modified scripts ### Reproduction The behavior of the "Model/Character Settings" dialog...
### Bug Report With Hermes 2 Pro Mistral 7B, in certain situations chat.exe causes high CPU usage even after generation has finished. ### Steps to Reproduce 1. Install [Hermes 2...
### Bug Report Whichever Python script I run, when calling the GPT4All() constructor, say like this: `model = GPT4All(model_name='openchat-3.6-8b-20240522-Q5_K_M.gguf', allow_download=False, device='cpu')` ...I get the following error messages: ``` Failed to...
### Bug Report When one enters a certain prompts (the text from https://users.ece.cmu.edu/~gamvrosi/thelastq.html for instance), the chat gets stuck: there is apparently some processing going on, but it never produces...
When trying to use Solar Pro Preview (Maziyar Panahi's Q5_K_M GGUF), I'm largely getting incoherent responses:  Whether this is a problem with the GGUF or the inference logic is...
### Feature Request Currently, when there is an error loading the model, the following explanation is provided: > Model loading failures can happen for a variety of reasons, but the...