llama icon indicating copy to clipboard operation
llama copied to clipboard

Inference code for Llama models

Results 412 llama issues
Sort by recently updated
recently updated
newest added

This change will improve the user experience by enabling them to easily experiment with their own prompts without any unnecessary setup.

CLA Signed

Hi, I want to load 13B or larger model in a single A100 80G, but find that the two shards of the model are ought to be loaded in 2...

It might be useful if you get the model to work to write down the model (e.g. 7B) and the hardware you got it to run on. Then people can...

Interested to see if anyone is able to run on google colab. Seems like 16 GB should be enough and is granted often for colab free. Not sure if Colab...

too many 4channers on here. ![download](https://user-images.githubusercontent.com/126925202/222951369-ae2099cb-4c43-4df1-9d4b-b90e4b860de0.jpg)

So are people with AMD GPU's screwed? I literally just sold my nvidia card and a Radeon two days ago. I've been trying my hardest to get this damn thing...

Hi all, This isn't my work, but the initial promise of this model for many people was that it could potentially be run on consumer hardware. https://github.com/tloen/llama-int8.git I found this...

I execute the command in README for unconditional generation and do not change any hyper-parameters in example.py. The prompt I use is "Michael Jackson was tried for child sexual abuse...

I was able to download the 7B weights on Mac OS Monterey. I get the following errors when I try to call the example from the README in my Terminal:...