ComfyUI icon indicating copy to clipboard operation
ComfyUI copied to clipboard

ComfyUI crashes instantly after clicking “Generate” on Windows 11 with ROCm + RX 6600 XT

Open hamzeht8 opened this issue 3 weeks ago • 7 comments

Custom Node Testing

Your question

I successfully installed the new ROCm-enabled PyTorch on Windows 11 and everything loads fine in ComfyUI — the GPU is detected, models load, and there are no errors during startup.

However, the moment I click Generate, ComfyUI immediately crashes with no error message (the terminal just closes). This only happens during inference; model loading works normally.

Here’s my setup and logs:

GPU: AMD Radeon RX 6600 XT (gfx1032)

Driver: latest AMD ROCm PyTorch driver

PyTorch: 2.9.0+rocmsdk20251116

ROCm runtime: 7.1.52802

OS: Windows 11 Pro

Python: 3.12 (Anaconda)

ComfyUI: v0.3.76

torch.cuda.is_available() → True

Device name is correctly detected

I also tested:

HSA_OVERRIDE_GFX_VERSION=10.3.0 → same crash

Has anyone with a 6600 XT on Windows + ROCm gotten ComfyUI inference to run without crashing? Is this a ROCm issue or a ComfyUI compatibility problem with RDNA2?

Any help or debugging tips would be appreciated.

Full logs :

(base) C:\Users\merfk>conda activate comfyui_rocm

(comfyui_rocm) C:\Users\merfk>python -c "import torch" 2>nul && echo Success || echo Failure

Success

(comfyui_rocm) C:\Users\merfk>python -c "import torch; print(torch.cuda.is_available())"

True

(comfyui_rocm) C:\Users\merfk>python -c "import torch; print(f'device name [0]:', torch.cuda.get_device_name(0))"

device name [0]: AMD Radeon RX 6600 XT

(comfyui_rocm) C:\Users\merfk>python -m torch.utils.collect_env

:128: RuntimeWarning: 'torch.utils.collect_env' found in sys.modules after import of package 'torch.utils', but prior to execution of 'torch.utils.collect_env'; this may result in unpredictable behaviour

Collecting environment information...

PyTorch version: 2.9.0+rocmsdk20251116

Is debug build: False

CUDA used to build PyTorch: N/A

ROCM used to build PyTorch: 7.1.52802-561cc400e1

OS: Microsoft Windows 11 Pro (10.0.26200 64-bit)

GCC version: Could not collect

Clang version: Could not collect

CMake version: Could not collect

Libc version: N/A

Python version: 3.12.12 | packaged by Anaconda, Inc. | (main, Oct 21 2025, 20:05:38) [MSC v.1929 64 bit (AMD64)] (64-bit runtime)

Python platform: Windows-11-10.0.26200-SP0

Is CUDA available: True

CUDA runtime version: Could not collect

CUDA_MODULE_LOADING set to:

GPU models and configuration: AMD Radeon RX 6600 XT (gfx1032)

Nvidia driver version: Could not collect

cuDNN version: Could not collect

Is XPU available: False

HIP runtime version: 7.1.52802

MIOpen runtime version: 3.5.1

Is XNNPACK available: True

CPU:

Name: Intel(R) Core(TM) i5-10400F CPU @ 2.90GHz

Manufacturer: GenuineIntel

Family: 205

Architecture: 9

ProcessorType: 3

DeviceID: CPU0

CurrentClockSpeed: 2904

MaxClockSpeed: 2904

L2CacheSize: 1536

L2CacheSpeed: None

Revision: None

Versions of relevant libraries:

[pip3] numpy==2.3.5

[pip3] torch==2.9.0+rocmsdk20251116

[pip3] torchaudio==2.9.0+rocmsdk20251116

[pip3] torchsde==0.2.6

[pip3] torchvision==0.24.0+rocmsdk20251116

[conda] numpy 2.3.5 pypi_0 pypi

[conda] torch 2.9.0+rocmsdk20251116 pypi_0 pypi

[conda] torchaudio 2.9.0+rocmsdk20251116 pypi_0 pypi

[conda] torchsde 0.2.6 pypi_0 pypi

[conda] torchvision 0.24.0+rocmsdk20251116 pypi_0 pypi

(comfyui_rocm) C:\Users\merfk>cd Comfyui

(comfyui_rocm) C:\Users\merfk\ComfyUI>

(comfyui_rocm) C:\Users\merfk\ComfyUI>set HSA_OVERRIDE_GFX_VERSION=10.3.0

(comfyui_rocm) C:\Users\merfk\ComfyUI>python main.py

Checkpoint files will always be loaded safely.

Total VRAM 8176 MB, total RAM 32667 MB

pytorch version: 2.9.0+rocmsdk20251116

Set: torch.backends.cudnn.enabled = False for better AMD performance.

AMD arch: gfx1032

ROCm version: (7, 1)

Set vram state to: NORMAL_VRAM

Device: cuda:0 AMD Radeon RX 6600 XT : native

Enabled pinned memory 14700.0

Using sub quadratic optimization for attention, if you have memory or speed issues try using: --use-split-cross-attention

Python version: 3.12.12 | packaged by Anaconda, Inc. | (main, Oct 21 2025, 20:05:38) [MSC v.1929 64 bit (AMD64)]

ComfyUI version: 0.3.76

ComfyUI frontend version: 1.33.10

[Prompt Server] web root: C:\Users\merfk.conda\envs\comfyui_rocm\Lib\site-packages\comfyui_frontend_package\static

Total VRAM 8176 MB, total RAM 32667 MB

pytorch version: 2.9.0+rocmsdk20251116

Set: torch.backends.cudnn.enabled = False for better AMD performance.

AMD arch: gfx1032

ROCm version: (7, 1)

Set vram state to: NORMAL_VRAM

Device: cuda:0 AMD Radeon RX 6600 XT : native

Enabled pinned memory 14700.0

Import times for custom nodes:

0.0 seconds: C:\Users\merfk\ComfyUI\custom_nodes\websocket_image_save.py

Context impl SQLiteImpl.

Will assume non-transactional DDL.

No target revision found.

Starting server

To see the GUI go to: http://127.0.0.1:8188/

got prompt

model weight dtype torch.float16, manual cast: None

model_type EPS

Using split attention in VAE

Using split attention in VAE

VAE load device: cuda:0, offload device: cpu, dtype: torch.bfloat16

Requested to load SDXLClipModel

loaded completely; 95367431640625005117571072.00 MB usable, 1560.80 MB loaded, full load: True

CLIP/text encoder model load device: cuda:0, offload device: cpu, current: cuda:0, dtype: torch.float16

Requested to load SDXLClipModel

(comfyui_rocm) C:\Users\merfk\ComfyUI>

Logs


Other

No response

hamzeht8 avatar Dec 03 '25 18:12 hamzeht8

The instant crash is usually related to the pagefile size. Try setting it "managed by the system" or increase the limit.

LukeG89 avatar Dec 03 '25 18:12 LukeG89

The instant crash is usually related to the pagefile size. Try setting it "managed by the system" or increase the limit.

Thanks for the suggestion, I tried increasing the pagefile size, but unfortunately the crash still happens immediately when I click Generate. (comfyui_rocm) C:\Users\merfk\ComfyUI>python main.py --use-split-cross-attention Checkpoint files will always be loaded safely. Total VRAM 8176 MB, total RAM 32667 MB pytorch version: 2.9.0+rocmsdk20251116 Set: torch.backends.cudnn.enabled = False for better AMD performance. AMD arch: gfx1032 ROCm version: (7, 1) Set vram state to: NORMAL_VRAM Device: cuda:0 AMD Radeon RX 6600 XT : native Enabled pinned memory 14700.0 Using split optimization for attention Python version: 3.12.12 | packaged by Anaconda, Inc. | (main, Oct 21 2025, 20:05:38) [MSC v.1929 64 bit (AMD64)] ComfyUI version: 0.3.76 ComfyUI frontend version: 1.33.10 [Prompt Server] web root: C:\Users\merfk.conda\envs\comfyui_rocm\Lib\site-packages\comfyui_frontend_package\static Total VRAM 8176 MB, total RAM 32667 MB pytorch version: 2.9.0+rocmsdk20251116 Set: torch.backends.cudnn.enabled = False for better AMD performance. AMD arch: gfx1032 ROCm version: (7, 1) Set vram state to: NORMAL_VRAM Device: cuda:0 AMD Radeon RX 6600 XT : native Enabled pinned memory 14700.0

Import times for custom nodes: 0.0 seconds: C:\Users\merfk\ComfyUI\custom_nodes\websocket_image_save.py

Context impl SQLiteImpl. Will assume non-transactional DDL. No target revision found. Starting server

To see the GUI go to: http://127.0.0.1:8188 got prompt model weight dtype torch.float16, manual cast: None model_type EPS Using split attention in VAE Using split attention in VAE VAE load device: cuda:0, offload device: cpu, dtype: torch.bfloat16 Requested to load SDXLClipModel loaded completely; 95367431640625005117571072.00 MB usable, 1560.80 MB loaded, full load: True CLIP/text encoder model load device: cuda:0, offload device: cpu, current: cuda:0, dtype: torch.float16 Requested to load SDXLClipModel

(comfyui_rocm) C:\Users\merfk\ComfyUI>

hamzeht8 avatar Dec 03 '25 20:12 hamzeht8

where did you get this rocm that has 6600xt or any rdna2 is recognised and working with in windows ?

patientx avatar Dec 03 '25 21:12 patientx

where did you get this rocm that has 6600xt or any rdna2 is recognised and working with in windows ? I used those guides https://www.amd.com/en/resources/support-articles/release-notes/RN-AMDGPU-WINDOWS-PYTORCH-7-1-1.html

https://rocm.docs.amd.com/projects/radeon-ryzen/en/latest/docs/install/installrad/windows/install-pytorch.html

And I saw a post in your reddit that said this:

just regular windows 11. pytorch instructions are here: https://github.com/ROCm/TheRock/tree/main/external-builds/pytorch#build-instructions . you should be able to use to use rocm wheels from here (need all 4): https://rocm.nightlies.amd.com/v2/gfx103X-dgpu/ pytorch took me ~2h to compile and you may need to install some of the dependencies from here: https://github.com/ROCm/TheRock/blob/main/docs/development/windows_support.md#building-therock-from-source . if rocm wheels are not compatible with your gpu, try using env. variable HSA_OVERRIDE_GFX_VERSION="10.3.0"

(I didn’t follow that guide Should I have followed that guide?)

hamzeht8 avatar Dec 04 '25 06:12 hamzeht8

I have tried compiling for a full day and always some problems ,as far as I know unless you do extensive patching and manual fixing they still do not work for 103x, there is some things happening though in their github

patientx avatar Dec 04 '25 11:12 patientx

I have tried compiling for a full day and always some problems ,as far as I know unless you do extensive patching and manual fixing they still do not work for 103x, there is some things happening though in their github

How long do you think it will take for ROCm on Windows to become fully stable? And do you think it will properly support the RX 6000 series, or should we just wait for newer ZLUDA releases instead?

hamzeht8 avatar Dec 04 '25 12:12 hamzeht8

I have tried compiling for a full day and always some problems ,as far as I know unless you do extensive patching and manual fixing they still do not work for 103x, there is some things happening though in their github

How long do you think it will take for ROCm on Windows to become fully stable? And do you think it will properly support the RX 6000 series, or should we just wait for newer ZLUDA releases instead?

they are getting there, follow this directory, when there is torch / torchvision / torchaudio added and it includes windows packages it would mean it is done .

https://rocm.nightlies.amd.com/v2/gfx103X-dgpu/

patientx avatar Dec 04 '25 12:12 patientx