blog
blog copied to clipboard
Public repo for HF blog posts
is there a library I could use to distribute model loading b/w gpu and cpu, I have gpu with 16gb memory and tried https://huggingface.co/blog/assisted-generation (the model upto 1.3b params works...
In the following paragraph of the blog post [SafeCoder vs Closed-source Code Assistants](https://huggingface.co/blog/safecoder-vs-closed-source-code-assistants) ` However, these services rely on closed-source models that can't be customized to your technical culture and...
Without this correction, the code will not run because the `labels` is passed as a kwargs inside `Trainer`.
I think I was added by accident to the author list as I didn't contribute anything to this release :) As a result, I suggest removing me from the author...
Kernel specifications: ``` Image: Data Science 3.0 Kernel: Python 3 Instance type: ml.t3.medium Start-up script: No script ``` This is my exact notebook code, copied from the "Deploy" button on...
Hi, It is not clear if we need to follow the prompt template for inference using pipeline as mentioned [here](https://github.com/huggingface/blog/blob/main/llama2.md#how-to-prompt-llama-2) or do we need to follow the pipeline code without...
hi, I translated the article into Persian language. Dr. Lambert told me to make a PR in Github reop so the rest of the work could be done. thanks P.S....
Hi Team, Some of our users are interested in launching and running Gorilla LLM models on sagemaker. I couldn't tell from [this](https://github.com/huggingface/blog/blob/main/sagemaker-huggingface-llm.md) blog if that was possible? An update would...
Tried following this in my sagemaker notebook instance (g5.48x.large). Unable to use this script for falcon-7b. Keep getting the error: Shard cannot start. The steps work well for falcon-40b and...