llama-gpt icon indicating copy to clipboard operation
llama-gpt copied to clipboard

how to download the load the model

Open jimleee opened this issue 2 years ago • 2 comments

i am trying to install with docker, i am now in this step:“llama-gpt-llama-gpt-ui-1 | [INFO wait] Host [llama-gpt-api-13b:8000] not yet available..”, now my question is how to download the load the model? below is the guideline.

Note: On the first run, it may take a while for the model to be downloaded to the /models directory. You may see lots of output like for a few minutes, which is normal:

llama-gpt-llama-gpt-ui-1 | [INFO wait] Host [llama-gpt-api-13b:8000] not yet available... After the model has been downloaded and loaded, and the API server is running, you'll see an output like:

llama-gpt-llama-gpt-api-13b-1 | INFO: Uvicorn running on http://0.0.0.0:8000 (Press CTRL+C to quit)

jimleee avatar Aug 22 '23 14:08 jimleee

How long does it take to download? What connection speeds to you have?

debarko avatar Aug 30 '23 09:08 debarko

@jimlee - it happens automatically upon starting the docker container and totally depends on your internet download speeds. See this for a sense of the size:

https://github.com/getumbrel/llama-gpt#supported-models

On Windows, I used the Task Manager > Performance tab to monitor the speed and completion. It will then take a while to start up afterwards. Note - if you happen to interrupt this while the download is underway, you may need to delete the model from /models/ and restart the containers to retry.

arch1v1st avatar Sep 24 '23 00:09 arch1v1st