localGPT
localGPT copied to clipboard
Unable to load llama model from path
I've been getting a repeated error when trying to run localgpt. I got everything working under cpu but the performance was pretty slow so I rebuilt using the latest cuda version of pytorch and rebuilt the llama wheel but now I can ingest the test document ok but when I go to run I get the following issue every time:
(localgpt) PS J:\localgpt\localGPT> python run_localGPT.py
2024-01-31 13:41:26,923 - INFO - run_localGPT.py:241 - Running on: cuda
2024-01-31 13:41:26,923 - INFO - run_localGPT.py:242 - Display Source Documents set to: False
2024-01-31 13:41:26,923 - INFO - run_localGPT.py:243 - Use history set to: False
2024-01-31 13:41:27,344 - INFO - SentenceTransformer.py:66 - Load pretrained SentenceTransformer: hkunlp/instructor-large
load INSTRUCTOR_Transformer
C:\ProgramData\anaconda3\envs\localgpt\lib\site-packages\torch_utils.py:831: UserWarning: TypedStorage is deprecated. It will be removed in the future and UntypedStorage will be the only storage class. This should only matter to you if you are using storages directly. To access UntypedStorage directly, use tensor.untyped_storage() instead of tensor.storage()
return self.fget.get(instance, owner)()
max_seq_length 512
2024-01-31 13:41:28,335 - INFO - run_localGPT.py:59 - Loading Model: TheBloke/Llama-2-7b-Chat-GGUF, on: cuda
2024-01-31 13:41:28,335 - INFO - run_localGPT.py:60 - This action can take a few minutes!
2024-01-31 13:41:28,335 - INFO - load_models.py:38 - Using Llamacpp for GGUF/GGML quantized models
Traceback (most recent call last):
File "J:\localgpt\localGPT\run_localGPT.py", line 282, in
I have checked my environment and can see the model folder has a symlink file to the model in the location that is listed in the error so assume that is ok and my pc paths don't seem to show anything mentioning the cuda folders with the repeated bin\bin so I'm not quite sure where its reading this from to get these errors, does anyone have any ideas?
How did you install llama-cpp-python? Make sure you install the version specified in the readme:
CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 pip install llama-cpp-python==0.1.83 --no-cache-dir
@PromtEngineer CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 pip install llama-cpp-python==0.1.83 --no-cache-dir
what if we don't Install the specified version what will happen?