Sean Owen
Sean Owen
Yes at the moment that is the substantial difference.
You need to choose a GPU runtime - and _unselect_ "Use Photon"
Closing as duplicate of #2
You need to select a 12.2 ML _GPU_ runtimes to see GPU instance types. (Also, and this bugs me, you need to unselect "Use Photon" to see them.)
I believe the issue is that the weights are a function of the training data, which has a non-commercial license and ultimately stems from OpenAI output. That means downstream users...
I'm not sure if we're going to edit the dataset unless it contains harmful or sensitive info (shouldn't!) or obvious content problems. In a way we want to capture the...
oh yeah fair, but I'll leave it to maintainers to decide policies on typos like that
@matthayes is going to collect this feedback and get it into a larger batch update, thank you
Can you try adding `task="text-generation"` to the `pipeline()`? That's worked for a few other people. @matthayes I think this task `instruction-following` thing might be causing this issue.
As a rule of thumb, because this model is stored in 16-bit, you will need 2 x num parameters bytes of mem, plus memory for the data and so on....