stable-diffusion-webui icon indicating copy to clipboard operation
stable-diffusion-webui copied to clipboard

[Bug]: AssertionError: Torch not compiled with CUDA enable

Open JohnieXu opened this issue 11 months ago • 6 comments

Checklist

  • [X] The issue exists after disabling all extensions
  • [X] The issue exists on a clean installation of webui
  • [x] The issue is caused by an extension, but I believe it is caused by a bug in the webui
  • [X] The issue exists in the current version of the webui
  • [X] The issue has not been reported before recently
  • [x] The issue has been reported before but has not been fixed yet

What happened?

when i use txt2img, an error happened: AssertionError: Torch not compiled with CUDA enable

Steps to reproduce the problem

Python 3.10.6 (main, Mar  9 2024, 12:41:50) [Clang 15.0.0 (clang-1500.0.40.1)]
Version: v1.8.0
Commit hash: bef51aed032c0aaa5cfd80445bc4cf0d85b408b5
Launching Web UI with arguments: --skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
Warning: caught exception 'Torch not compiled with CUDA enabled', memory monitor disabled
==============================================================================
You are running torch 2.1.0.
The program is tested to work with torch 2.1.2.
To reinstall the desired version, run with commandline flag --reinstall-torch.
Beware that this will cause a lot of large files to be downloaded, as well as
there are reports of issues with training tab on the latest version.

Use --skip-version-check commandline argument to disable this check.
==============================================================================
Loading weights [6ce0161689] from /Volumes/RTL9210/SD/stable-diffusion-webui/models/Stable-diffusion/v1-5-pruned-emaonly.safetensors
Creating model from config: /Volumes/RTL9210/SD/stable-diffusion-webui/configs/v1-inference.yaml
Running on local URL:  http://127.0.0.1:7860

What should have happened?

no error happened

What browsers do you use to access the UI ?

Google Chrome

Sysinfo

{ "Platform": "macOS-13.6.3-x86_64-i386-64bit", "Python": "3.10.6", "Version": "v1.8.0", "Commit": "bef51aed032c0aaa5cfd80445bc4cf0d85b408b5", "Script path": "/Volumes/RTL9210/SD/stable-diffusion-webui", "Data path": "/Volumes/RTL9210/SD/stable-diffusion-webui", "Extensions dir": "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions", "Checksum": "416a4ccbaf45286a314038b0a5e9334f25882bc163eedaff19b4b210c8053194", "Commandline": [ "launch.py", "--skip-torch-cuda-test", "--upcast-sampling", "--no-half-vae", "--use-cpu", "interrogate" ], "Torch env info": { "torch_version": "2.1.0", "is_debug_build": "False", "cuda_compiled_version": null, "gcc_version": null, "clang_version": "15.0.0 (clang-1500.0.40.1)", "cmake_version": "version 3.28.1", "os": "macOS 13.6.3 (x86_64)", "libc_version": "N/A", "python_version": "3.10.6 (main, Mar 9 2024, 12:41:50) [Clang 15.0.0 (clang-1500.0.40.1)] (64-bit runtime)", "python_platform": "macOS-13.6.3-x86_64-i386-64bit", "is_cuda_available": "False", "cuda_runtime_version": null, "cuda_module_loading": "N/A", "nvidia_driver_version": null, "nvidia_gpu_models": null, "cudnn_version": null, "pip_version": "pip3", "pip_packages": [ "numpy==1.26.2", "open-clip-torch==2.20.0", "pytorch-lightning==1.9.4", "torch==2.1.0", "torchdiffeq==0.2.3", "torchmetrics==1.3.1", "torchsde==0.2.6", "torchvision==0.16.0" ], "conda_packages": null, "hip_compiled_version": "N/A", "hip_runtime_version": "N/A", "miopen_runtime_version": "N/A", "caching_allocator_config": "", "is_xnnpack_available": "True", "cpu_info": "Intel(R) Core(TM) i5-7360U CPU @ 2.30GHz" }, "Exceptions": [ { "exception": "Torch not compiled with CUDA enabled", "traceback": [ [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py, line 57, f", "res = list(func(*args, **kwargs))" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py, line 36, f", "res = func(*args, **kwargs)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/txt2img.py, line 110, txt2img", "processed = processing.process_images(p)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py, line 785, process_images", "res = process_images_inner(p)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py, line 854, process_images_inner", "with devices.autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 218, autocast", "if has_xpu() or has_mps() or cuda_no_autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 28, cuda_no_autocast", "device_id = get_cuda_device_id()" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 40, get_cuda_device_id", ") or torch.cuda.current_device()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 769, current_device", "_lazy_init()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 289, _lazy_init", "raise AssertionError("Torch not compiled with CUDA enabled")" ] ] }, { "exception": "Torch not compiled with CUDA enabled", "traceback": [ [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py, line 57, f", "res = list(func(*args, **kwargs))" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py, line 36, f", "res = func(*args, **kwargs)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/txt2img.py, line 110, txt2img", "processed = processing.process_images(p)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py, line 785, process_images", "res = process_images_inner(p)" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py, line 854, process_images_inner", "with devices.autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 218, autocast", "if has_xpu() or has_mps() or cuda_no_autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 28, cuda_no_autocast", "device_id = get_cuda_device_id()" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 40, get_cuda_device_id", ") or torch.cuda.current_device()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 769, current_device", "_lazy_init()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 289, _lazy_init", "raise AssertionError("Torch not compiled with CUDA enabled")" ] ] }, { "exception": "Torch not compiled with CUDA enabled", "traceback": [ [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py, line 621, get_sd_model", "load_model()" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py, line 771, load_model", "with devices.autocast(), torch.no_grad():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 218, autocast", "if has_xpu() or has_mps() or cuda_no_autocast():" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 28, cuda_no_autocast", "device_id = get_cuda_device_id()" ], [ "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py, line 40, get_cuda_device_id", ") or torch.cuda.current_device()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 769, current_device", "_lazy_init()" ], [ "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/init.py, line 289, _lazy_init", "raise AssertionError("Torch not compiled with CUDA enabled")" ] ] } ], "CPU": { "model": "i386", "count logical": 4, "count physical": 2 }, "RAM": { "total": "16GB", "used": "9GB", "free": "72MB", "active": "5GB", "inactive": "5GB" }, "Extensions": [], "Inactive extensions": [], "Environment": { "COMMANDLINE_ARGS": "--skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate", "GIT": "git", "GRADIO_ANALYTICS_ENABLED": "False", "TORCH_COMMAND": "pip install torch==2.1.0 torchvision==0.16.0" }, "Config": { "ldsr_steps": 100, "ldsr_cached": false, "SCUNET_tile": 256, "SCUNET_tile_overlap": 8, "SWIN_tile": 192, "SWIN_tile_overlap": 8, "SWIN_torch_compile": false, "hypertile_enable_unet": false, "hypertile_enable_unet_secondpass": false, "hypertile_max_depth_unet": 3, "hypertile_max_tile_unet": 256, "hypertile_swap_size_unet": 3, "hypertile_enable_vae": false, "hypertile_max_depth_vae": 3, "hypertile_max_tile_vae": 128, "hypertile_swap_size_vae": 3, "sd_model_checkpoint": "v1-5-pruned-emaonly.safetensors [6ce0161689]", "sd_checkpoint_hash": "6ce0161689b3853acaa03779ec93eafe75a02f4ced659bee03f50797806fa2fa" }, "Startup": { "total": 11.264039754867554, "records": { "initial startup": 0.01104283332824707, "prepare environment/checks": 7.796287536621094e-05, "prepare environment/git version info": 0.04249715805053711, "prepare environment/torch GPU test": 0.0039539337158203125, "prepare environment/clone repositores": 0.13756275177001953, "prepare environment/run extensions installers": 0.00048422813415527344, "prepare environment": 0.2236621379852295, "launcher": 0.0011608600616455078, "import torch": 4.559276103973389, "import gradio": 1.2403457164764404, "setup paths": 0.965001106262207, "import ldm": 0.003961086273193359, "import sgm": 5.0067901611328125e-06, "initialize shared": 0.14409089088439941, "other imports": 0.9297831058502197, "opts onchange": 0.0005729198455810547, "setup SD model": 9.393692016601562e-05, "setup codeformer": 0.0010669231414794922, "setup gfpgan": 0.006997108459472656, "set samplers": 4.601478576660156e-05, "list extensions": 0.0011892318725585938, "restore config state file": 1.0013580322265625e-05, "list SD models": 0.08728671073913574, "list localizations": 0.000331878662109375, "load scripts/custom_code.py": 0.0021550655364990234, "load scripts/img2imgalt.py": 0.00045371055603027344, "load scripts/loopback.py": 0.0002651214599609375, "load scripts/outpainting_mk_2.py": 0.0002989768981933594, "load scripts/poor_mans_outpainting.py": 0.00023317337036132812, "load scripts/postprocessing_caption.py": 0.00018787384033203125, "load scripts/postprocessing_codeformer.py": 0.00023293495178222656, "load scripts/postprocessing_create_flipped_copies.py": 0.00018787384033203125, "load scripts/postprocessing_focal_crop.py": 0.0006661415100097656, "load scripts/postprocessing_gfpgan.py": 0.00019788742065429688, "load scripts/postprocessing_split_oversized.py": 0.00019931793212890625, "load scripts/postprocessing_upscale.py": 0.0002818107604980469, "load scripts/processing_autosized_crop.py": 0.00021123886108398438, "load scripts/prompt_matrix.py": 0.0002276897430419922, "load scripts/prompts_from_file.py": 0.0003600120544433594, "load scripts/sd_upscale.py": 0.00022125244140625, "load scripts/xyz_grid.py": 0.002402782440185547, "load scripts/ldsr_model.py": 0.5824201107025146, "load scripts/lora_script.py": 0.1882781982421875, "load scripts/scunet_model.py": 0.030708789825439453, "load scripts/swinir_model.py": 0.03022289276123047, "load scripts/hotkey_config.py": 0.00026297569274902344, "load scripts/extra_options_section.py": 0.00027632713317871094, "load scripts/hypertile_script.py": 0.05973458290100098, "load scripts/hypertile_xyz.py": 0.00025534629821777344, "load scripts/soft_inpainting.py": 0.0003829002380371094, "load scripts/comments.py": 0.029378890991210938, "load scripts/refiner.py": 0.0002911090850830078, "load scripts/seed.py": 0.00023698806762695312, "load scripts": 0.9312648773193359, "load upscalers": 0.0019369125366210938, "refresh VAE": 0.0017819404602050781, "refresh textual inversion templates": 7.605552673339844e-05, "scripts list_optimizers": 0.00040721893310546875, "scripts list_unets": 6.9141387939453125e-06, "reload hypernetworks": 0.21456289291381836, "initialize extra networks": 0.03221607208251953, "scripts before_ui_callback": 0.0063779354095458984, "create ui": 1.0929930210113525, "gradio launch": 0.7930488586425781, "add APIs": 0.05200815200805664, "app_started_callback/lora_script.py": 0.0007350444793701172, "app_started_callback": 0.0007450580596923828 } }, "Packages": [ "accelerate==0.21.0", "aenum==3.1.15", "aiofiles==23.2.1", "aiohttp==3.9.3", "aiosignal==1.3.1", "altair==5.2.0", "antlr4-python3-runtime==4.9.3", "anyio==3.7.1", "async-timeout==4.0.3", "attrs==23.2.0", "blendmodes==2022", "certifi==2024.2.2", "charset-normalizer==3.3.2", "clean-fid==0.1.35", "click==8.1.7", "clip==1.0", "contourpy==1.2.0", "cycler==0.12.1", "deprecation==2.1.0", "einops==0.4.1", "exceptiongroup==1.2.0", "facexlib==0.3.0", "fastapi==0.94.0", "ffmpy==0.3.2", "filelock==3.13.1", "filterpy==1.4.5", "fonttools==4.49.0", "frozenlist==1.4.1", "fsspec==2024.2.0", "ftfy==6.1.3", "gitdb==4.0.11", "gitpython==3.1.32", "gradio-client==0.5.0", "gradio==3.41.2", "h11==0.12.0", "httpcore==0.15.0", "httpx==0.24.1", "huggingface-hub==0.21.4", "idna==3.6", "imageio==2.34.0", "importlib-resources==6.1.3", "inflection==0.5.1", "jinja2==3.1.3", "jsonmerge==1.8.0", "jsonschema-specifications==2023.12.1", "jsonschema==4.21.1", "kiwisolver==1.4.5", "kornia==0.6.7", "lark==1.1.2", "lazy-loader==0.3", "lightning-utilities==0.10.1", "llvmlite==0.42.0", "markupsafe==2.1.5", "matplotlib==3.8.3", "mpmath==1.3.0", "multidict==6.0.5", "networkx==3.2.1", "numba==0.59.0", "numpy==1.26.2", "omegaconf==2.2.3", "open-clip-torch==2.20.0", "opencv-python==4.9.0.80", "orjson==3.9.15", "packaging==23.2", "pandas==2.2.1", "piexif==1.1.3", "pillow==9.5.0", "pip==22.2.1", "protobuf==3.20.0", "psutil==5.9.5", "pydantic==1.10.14", "pydub==0.25.1", "pyparsing==3.1.2", "python-dateutil==2.9.0.post0", "python-multipart==0.0.9", "pytorch-lightning==1.9.4", "pytz==2024.1", "pywavelets==1.5.0", "pyyaml==6.0.1", "referencing==0.33.0", "regex==2023.12.25", "requests==2.31.0", "resize-right==0.0.2", "rpds-py==0.18.0", "safetensors==0.4.2", "scikit-image==0.21.0", "scipy==1.12.0", "semantic-version==2.10.0", "sentencepiece==0.2.0", "setuptools==63.2.0", "six==1.16.0", "smmap==5.0.1", "sniffio==1.3.1", "spandrel==0.1.6", "starlette==0.26.1", "sympy==1.12", "tifffile==2024.2.12", "timm==0.9.16", "tokenizers==0.13.3", "tomesd==0.1.3", "toolz==0.12.1", "torch==2.1.0", "torchdiffeq==0.2.3", "torchmetrics==1.3.1", "torchsde==0.2.6", "torchvision==0.16.0", "tqdm==4.66.2", "trampoline==0.1.2", "transformers==4.30.2", "typing-extensions==4.10.0", "tzdata==2024.1", "urllib3==2.2.1", "uvicorn==0.27.1", "wcwidth==0.2.13", "websockets==11.0.3", "yarl==1.9.4" ] }

Console logs

stable-diffusion-webui git:(master) ./webui.sh

################################################################
Install script for stable-diffusion + Web UI
Tested on Debian 11 (Bullseye), Fedora 34+ and openSUSE Leap 15.4 or newer.
################################################################

################################################################
Running on xuzhiyong user
################################################################

################################################################
Repo already cloned, using it as install directory
################################################################

################################################################
Create and activate python venv
################################################################

################################################################
Launching launch.py...
################################################################
Python 3.10.6 (main, Mar  9 2024, 12:41:50) [Clang 15.0.0 (clang-1500.0.40.1)]
Version: v1.8.0
Commit hash: bef51aed032c0aaa5cfd80445bc4cf0d85b408b5
Launching Web UI with arguments: --skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
Warning: caught exception 'Torch not compiled with CUDA enabled', memory monitor disabled
==============================================================================
You are running torch 2.1.0.
The program is tested to work with torch 2.1.2.
To reinstall the desired version, run with commandline flag --reinstall-torch.
Beware that this will cause a lot of large files to be downloaded, as well as
there are reports of issues with training tab on the latest version.

Use --skip-version-check commandline argument to disable this check.
==============================================================================
Loading weights [6ce0161689] from /Volumes/RTL9210/SD/stable-diffusion-webui/models/Stable-diffusion/v1-5-pruned-emaonly.safetensors
Creating model from config: /Volumes/RTL9210/SD/stable-diffusion-webui/configs/v1-inference.yaml
Running on local URL:  http://127.0.0.1:7860

To create a public link, set `share=True` in `launch()`.
Startup time: 11.3s (prepare environment: 0.2s, import torch: 4.6s, import gradio: 1.2s, setup paths: 1.0s, initialize shared: 0.1s, other imports: 0.9s, load scripts: 0.9s, reload hypernetworks: 0.2s, create ui: 1.1s, gradio launch: 0.8s).
OMP: Warning #191: Forking a process while a parallel region is active is potentially unsafe.
OMP: Warning #191: Forking a process while a parallel region is active is potentially unsafe.
OMP: Warning #191: Forking a process while a parallel region is active is potentially unsafe.
OMP: Warning #191: Forking a process while a parallel region is active is potentially unsafe.
OMP: Warning #191: Forking a process while a parallel region is active is potentially unsafe.
OMP: Warning #191: Forking a process while a parallel region is active is potentially unsafe.
Applying attention optimization: InvokeAI... done.
loading stable diffusion model: AssertionError
Traceback (most recent call last):
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 973, in _bootstrap
    self._bootstrap_inner()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 953, in run
    self._target(*self._args, **self._kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/initialize.py", line 148, in load_model
    shared.sd_model  # noqa: B018
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/shared_items.py", line 133, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py", line 621, in get_sd_model
    load_model()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py", line 771, in load_model
    with devices.autocast(), torch.no_grad():
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 218, in autocast
    if has_xpu() or has_mps() or cuda_no_autocast():
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 28, in cuda_no_autocast
    device_id = get_cuda_device_id()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 40, in get_cuda_device_id
    ) or torch.cuda.current_device()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 769, in current_device
    _lazy_init()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 289, in _lazy_init
    raise AssertionError("Torch not compiled with CUDA enabled")
AssertionError: Torch not compiled with CUDA enabled


Stable diffusion model failed to load
Using already loaded model v1-5-pruned-emaonly.safetensors [6ce0161689]: done in 0.0s
*** Error completing request
*** Arguments: ('task(6gssuckqcqqbe6l)', <gradio.routes.Request object at 0x15305ae30>, 'AssertionError: Torch not compiled with CUDA enabled', '', [], 20, 'DPM++ 2M Karras', 1, 1, 7, 512, 512, False, 0.7, 2, 'Latent', 0, 0, 0, 'Use same checkpoint', 'Use same sampler', '', '', [], 0, False, '', 0.8, -1, False, -1, 0, 0, 0, False, False, 'positive', 'comma', 0, False, False, 'start', '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, False, False, False, 0, False) {}
    Traceback (most recent call last):
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py", line 57, in f
        res = list(func(*args, **kwargs))
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py", line 36, in f
        res = func(*args, **kwargs)
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/txt2img.py", line 110, in txt2img
        processed = processing.process_images(p)
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py", line 785, in process_images
        res = process_images_inner(p)
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py", line 854, in process_images_inner
        with devices.autocast():
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 218, in autocast
        if has_xpu() or has_mps() or cuda_no_autocast():
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 28, in cuda_no_autocast
        device_id = get_cuda_device_id()
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 40, in get_cuda_device_id
        ) or torch.cuda.current_device()
      File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 769, in current_device
        _lazy_init()
      File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 289, in _lazy_init
        raise AssertionError("Torch not compiled with CUDA enabled")
    AssertionError: Torch not compiled with CUDA enabled

---
Exception in thread Thread-2 (load_model):
Traceback (most recent call last):
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 953, in run
    self._target(*self._args, **self._kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/initialize.py", line 153, in load_model
    devices.first_time_calculation()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 267, in first_time_calculation
    linear(x)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
    return forward_call(*args, **kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/networks.py", line 500, in network_Linear_forward
    return originals.Linear_forward(self, input)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/linear.py", line 114, in forward
    return F.linear(input, self.weight, self.bias)
RuntimeError: "addmm_impl_cpu_" not implemented for 'Half'
*** Error completing request
*** Arguments: ('task(2l51bhlon8pmei9)', <gradio.routes.Request object at 0x15305abf0>, 'AssertionError: Torch not compiled with CUDA enabled', '', [], 20, 'DPM++ 2M Karras', 1, 1, 7, 512, 512, False, 0.7, 2, 'Latent', 0, 0, 0, 'Use same checkpoint', 'Use same sampler', '', '', [], 0, False, '', 0.8, -1, False, -1, 0, 0, 0, False, False, 'positive', 'comma', 0, False, False, 'start', '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, False, False, False, 0, False) {}
    Traceback (most recent call last):
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py", line 57, in f
        res = list(func(*args, **kwargs))
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/call_queue.py", line 36, in f
        res = func(*args, **kwargs)
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/txt2img.py", line 110, in txt2img
        processed = processing.process_images(p)
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py", line 785, in process_images
        res = process_images_inner(p)
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/processing.py", line 854, in process_images_inner
        with devices.autocast():
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 218, in autocast
        if has_xpu() or has_mps() or cuda_no_autocast():
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 28, in cuda_no_autocast
        device_id = get_cuda_device_id()
      File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 40, in get_cuda_device_id
        ) or torch.cuda.current_device()
      File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 769, in current_device
        _lazy_init()
      File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 289, in _lazy_init
        raise AssertionError("Torch not compiled with CUDA enabled")
    AssertionError: Torch not compiled with CUDA enabled

---

Additional information

No response

JohnieXu avatar Mar 09 '24 11:03 JohnieXu

+1

huanglongyu avatar Mar 26 '24 03:03 huanglongyu

try adding --reinstall-torch to the commandline ARGS of webui. That helped me. Remember to remove it after reinstallation or it will reinstall every startup.

Xavia1991 avatar Mar 28 '24 08:03 Xavia1991

try adding --reinstall-torch to the commandline ARGS of webui. That helped me. Remember to remove it after reinstallation or it will reinstall every startup.

Thank u, but it doesn't work for me

################################################################
Install script for stable-diffusion + Web UI
Tested on Debian 11 (Bullseye), Fedora 34+ and openSUSE Leap 15.4 or newer.
################################################################

################################################################
Running on xuzhiyong user
################################################################

################################################################
Repo already cloned, using it as install directory
################################################################

################################################################
Create and activate python venv
################################################################

################################################################
Launching launch.py...
################################################################
Python 3.10.6 (main, Mar  9 2024, 12:41:50) [Clang 15.0.0 (clang-1500.0.40.1)]
Version: v1.8.0
Commit hash: bef51aed032c0aaa5cfd80445bc4cf0d85b408b5
Installing torch and torchvision
Looking in indexes: https://pypi.tuna.tsinghua.edu.cn/simple
Requirement already satisfied: torch==2.1.0 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (2.1.0)
Requirement already satisfied: torchvision==0.16.0 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (0.16.0)
Requirement already satisfied: typing-extensions in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (4.10.0)
Requirement already satisfied: sympy in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (1.12)
Requirement already satisfied: jinja2 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (3.1.3)
Requirement already satisfied: filelock in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (3.13.1)
Requirement already satisfied: fsspec in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (2024.2.0)
Requirement already satisfied: networkx in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torch==2.1.0) (3.2.1)
Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torchvision==0.16.0) (9.5.0)
Requirement already satisfied: numpy in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torchvision==0.16.0) (1.26.2)
Requirement already satisfied: requests in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from torchvision==0.16.0) (2.31.0)
Requirement already satisfied: MarkupSafe>=2.0 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from jinja2->torch==2.1.0) (2.1.5)
Requirement already satisfied: urllib3<3,>=1.21.1 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from requests->torchvision==0.16.0) (2.2.1)
Requirement already satisfied: charset-normalizer<4,>=2 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from requests->torchvision==0.16.0) (3.3.2)
Requirement already satisfied: idna<4,>=2.5 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from requests->torchvision==0.16.0) (3.6)
Requirement already satisfied: certifi>=2017.4.17 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from requests->torchvision==0.16.0) (2024.2.2)
Requirement already satisfied: mpmath>=0.19 in /Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages (from sympy->torch==2.1.0) (1.3.0)

[notice] A new release of pip available: 22.2.1 -> 24.0
[notice] To update, run: pip install --upgrade pip
Launching Web UI with arguments: --skip-torch-cuda-test --upcast-sampling --no-half-vae --use-cpu interrogate --reinstall-torch
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
Warning: caught exception 'Torch not compiled with CUDA enabled', memory monitor disabled
==============================================================================
You are running torch 2.1.0.
The program is tested to work with torch 2.1.2.
To reinstall the desired version, run with commandline flag --reinstall-torch.
Beware that this will cause a lot of large files to be downloaded, as well as
there are reports of issues with training tab on the latest version.

Use --skip-version-check commandline argument to disable this check.
==============================================================================
Loading weights [d124fc18f0] from /Volumes/RTL9210/SD/stable-diffusion-webui/models/Stable-diffusion/AOM3_orangemixs.safetensors
Creating model from config: /Volumes/RTL9210/SD/stable-diffusion-webui/configs/v1-inference.yaml
Running on local URL:  http://127.0.0.1:7860

To create a public link, set `share=True` in `launch()`.
Startup time: 19.0s (prepare environment: 3.5s, import torch: 7.2s, import gradio: 1.9s, setup paths: 2.2s, initialize shared: 0.2s, other imports: 1.1s, load scripts: 0.9s, reload hypernetworks: 0.2s, create ui: 0.9s, gradio launch: 0.7s).
Applying attention optimization: InvokeAI... done.
loading stable diffusion model: AssertionError
Traceback (most recent call last):
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 973, in _bootstrap
    self._bootstrap_inner()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 953, in run
    self._target(*self._args, **self._kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/initialize.py", line 148, in load_model
    shared.sd_model  # noqa: B018
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/shared_items.py", line 133, in sd_model
    return modules.sd_models.model_data.get_sd_model()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py", line 621, in get_sd_model
    load_model()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/sd_models.py", line 771, in load_model
    with devices.autocast(), torch.no_grad():
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 218, in autocast
    if has_xpu() or has_mps() or cuda_no_autocast():
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 28, in cuda_no_autocast
    device_id = get_cuda_device_id()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 40, in get_cuda_device_id
    ) or torch.cuda.current_device()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 769, in current_device
    _lazy_init()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/cuda/__init__.py", line 289, in _lazy_init
    raise AssertionError("Torch not compiled with CUDA enabled")
AssertionError: Torch not compiled with CUDA enabled


Stable diffusion model failed to load
Exception in thread Thread-2 (load_model):
Traceback (most recent call last):
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 1016, in _bootstrap_inner
    self.run()
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/threading.py", line 953, in run
    self._target(*self._args, **self._kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/initialize.py", line 153, in load_model
    devices.first_time_calculation()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/devices.py", line 267, in first_time_calculation
    linear(x)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1518, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1527, in _call_impl
    return forward_call(*args, **kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/networks.py", line 500, in network_Linear_forward
    return originals.Linear_forward(self, input)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/torch/nn/modules/linear.py", line 114, in forward
    return F.linear(input, self.weight, self.bias)
RuntimeError: "addmm_impl_cpu_" not implemented for 'Half'
Traceback (most recent call last):
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/routes.py", line 488, in run_predict
    output = await app.get_blocks().process_api(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/blocks.py", line 1431, in process_api
    result = await self.call_function(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/blocks.py", line 1103, in call_function
    prediction = await anyio.to_thread.run_sync(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/to_thread.py", line 33, in run_sync
    return await get_asynclib().run_sync_in_worker_thread(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
    return await future
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 703, in pages_html
    create_html()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 699, in create_html
    ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 699, in <listcomp>
    ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 518, in create_html
    self.items = {x["name"]: x for x in items_list}
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 518, in <dictcomp>
    self.items = {x["name"]: x for x in items_list}
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/ui_extra_networks_lora.py", line 82, in list_items
    item = self.create_item(name, index)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/ui_extra_networks_lora.py", line 69, in create_item
    elif shared.sd_model.is_sdxl and sd_version != network.SdVersion.SDXL:
AttributeError: 'NoneType' object has no attribute 'is_sdxl'
Traceback (most recent call last):
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/routes.py", line 488, in run_predict
    output = await app.get_blocks().process_api(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/blocks.py", line 1431, in process_api
    result = await self.call_function(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/blocks.py", line 1103, in call_function
    prediction = await anyio.to_thread.run_sync(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/to_thread.py", line 33, in run_sync
    return await get_asynclib().run_sync_in_worker_thread(
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 877, in run_sync_in_worker_thread
    return await future
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 807, in run
    result = context.run(func, *args)
  File "/Users/johniexu/.pyenv/versions/3.10.6/lib/python3.10/site-packages/gradio/utils.py", line 707, in wrapper
    response = f(*args, **kwargs)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 703, in pages_html
    create_html()
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 699, in create_html
    ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 699, in <listcomp>
    ui.pages_contents = [pg.create_html(ui.tabname) for pg in ui.stored_extra_pages]
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 518, in create_html
    self.items = {x["name"]: x for x in items_list}
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/modules/ui_extra_networks.py", line 518, in <dictcomp>
    self.items = {x["name"]: x for x in items_list}
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/ui_extra_networks_lora.py", line 82, in list_items
    item = self.create_item(name, index)
  File "/Volumes/RTL9210/SD/stable-diffusion-webui/extensions-builtin/Lora/ui_extra_networks_lora.py", line 69, in create_item
    elif shared.sd_model.is_sdxl and sd_version != network.SdVersion.SDXL:
AttributeError: 'NoneType' object has no attribute 'is_sdxl'

JohnieXu avatar Mar 29 '24 08:03 JohnieXu