text-generation-webui
text-generation-webui copied to clipboard
Support GLM-10b Model?
Here: https://huggingface.co/THUDM/glm-10b
They have some external code that has to run, which I enabled but it still errors. Is this a different type of model that can't load by CasualLLM?
I only ask because it's 9-11b is the sweet spot for 24gb without 8bit or any other tricks.
There is also https://huggingface.co/hyunwoongko/reddit-9B which looks like it would make a great chatbot but uses a different pytorch function. I can get it to load but not generate. Their model card certainly wins me over.
Correct me if I am wrong, but it seems to be a model for filling gaps like BERT rather than a model that predicts the next token for open-ended generation.
@oobabooga It's trained to do both. At least the 130b version is.
https://github.com/THUDM/ChatGLM-6B seems they have a model for chat now
Honestly ... even at 6b this model is impressive... far better than pygmalion....
This issue has been closed due to inactivity for 6 weeks. If you believe it is still relevant, please leave a comment below. You can tag a developer in your comment.