AutoAWQ icon indicating copy to clipboard operation
AutoAWQ copied to clipboard

GPU not detected

Open LucisVivae opened this issue 1 year ago • 1 comments

having issues with autoAWQ not detecting my RTX 3070ti with Cuda 12.4 installed when attempting to run with loading "TheBloke/Mistral-7B-OpenOrca-AWQ"

the command nvidia-smi returns:

Sun Apr  7 09:39:24 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 551.86                 Driver Version: 551.86         CUDA Version: 12.4     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                     TCC/WDDM  | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce RTX 3070 Ti   WDDM  |   00000000:01:00.0  On |                  N/A |
|  0%   43C    P8             18W /  290W |     571MiB /   8192MiB |      4%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

The 3 specific error messages are:

ValueError: Could not load model TheBloke/Mistral-7B-OpenOrca-AWQ with any of the following classes: (<class 'transformers.models.auto.modeling_auto.AutoModelForCausalLM'>, <class 'transformers.models.mistral.modeling_mistral.MistralForCausalLM'>). See the original errors:

ValueError: AWQ is only available on GPU

ValueError: AWQ is only available on GPU

LucisVivae avatar Apr 07 '24 00:04 LucisVivae

have got past the above by reinstalling torch which for some reason did not initially install cuda support. but now it is telling me "You have loaded an AWQ model on CPU and have a CUDA device available, make sure to set your model on a GPU device in order to run your model."

LucisVivae avatar Apr 07 '24 01:04 LucisVivae