localGPT
localGPT copied to clipboard
Requested tokens exceed context window of 4096
So when I run the localGPT is fine. but when I try to run_localGPT_API when I try to ask question it's always throw exceed context window error is there anyone how to solve this issue ?
Change it to a model that supports 8k or 16k tokens such as zephyr or Yi series. Also you will need to change the max tokens here
Thanks, I already try to change the model , but sometimes the answer is blank. Do you have any recommend local model that support 8k or 16k because when I find the zephyr and Yi series the GGUF one, the answer is blank.
@PromtEngineer Wouldn't changing the models will result into more hardware resources being consumed? Right? @fenry46 @PromtEngineer
Tell me more about it, please.