Mark Schmidt

Results 95 comments of Mark Schmidt

### Potential Solution 3 Simply run a cron job to scrape for updates from the [XML pages](https://rfc2822.gitlab.io/fdroid-firefox/fdroid/repo/index.xml) for a list of repos and analyze everything in a queue. The amount...

Duplicate of #1383

Put it in a folder next to the rest of your models and then select it from the dropdown. ![image](https://user-images.githubusercontent.com/5949853/233243705-feb1581b-612e-446c-8cfc-d67ecc74fdc7.png) Note that the "Tuned" versions need to be prompted like:...

> I think it needs to be quantized to run at full context in 24gb. There are some 16bit versions here: [huggingface.co/vvsotnikov/stablelm-tuned-alpha-7b-16bit](https://huggingface.co/vvsotnikov/stablelm-tuned-alpha-7b-16bit) [huggingface.co/vvsotnikov/stablelm-tuned-alpha-3b-16bit](https://huggingface.co/vvsotnikov/stablelm-tuned-alpha-3b-16bit) These might work with load-in-8bit aswell? (Haven't...

Yes the quality / benchmarks are currently not good, to say the least. I would hold out for the next training checkpoint tbh.

It's definitely not your PR's fault. See https://github.com/Stability-AI/StableLM/issues/30

I added a table of contents, cleaned up some confusing language, and made the "tags" full words for better readability ("bare" not "B" for example).

It's your thing man, go ahead. :)

> Good Job! I also want to try running on Colab Pro You can make the changes above to the Colab linked in the readme. https://colab.research.google.com/github/oobabooga/AI-Notebooks/blob/main/Colab-TextGen-GPU.ipynb Just click "Show Code"...

Nothing needs to be changed. Use load_in_8bit depending on how much VRAM your instance has and what size model you use. Refer to the table below. Model | 16bit VRAM...