Given groups=1, weight of size [320, 4, 3, 3], expected input[2, 16, 192, 128] to have 4 channels, but got 16 channels instead
Expected Behavior
Given groups=1, weight of size [320, 4, 3, 3], expected input[2, 16, 192, 128] to have 4 channels, but got 16 channels instead
Actual Behavior
Given groups=1, weight of size [320, 4, 3, 3], expected input[2, 16, 192, 128] to have 4 channels, but got 16 channels instead
Steps to Reproduce
Given groups=1, weight of size [320, 4, 3, 3], expected input[2, 16, 192, 128] to have 4 channels, but got 16 channels instead
Debug Logs
got prompt
model weight dtype torch.float16, manual cast: None
model_type EPS
Using pytorch attention in VAE
Using pytorch attention in VAE
Requested to load SD1ClipModel
Loading 1 new model
loaded completely 0.0 235.84423828125 True
Using pytorch attention in VAE
Using pytorch attention in VAE
final text_encoder_type: bert-base-uncased
Loads SAM model: D:\AI\ComfyUI\models\sams\sam_vit_b_01ec64.pth (device:Prefer GPU)
# of Detected SEGS: 1
1girl, solo, long_hair, looking_at_viewer, skirt, brown_hair, shirt, long_sleeves, brown_eyes, jewelry, standing, full_body, earrings, outdoors, shoes, day, socks, off_shoulder, nail_polish, black_footwear, sweater, tree, hand_on_hip, sleeves_past_wrists, plaid, window, leaf, plaid_skirt, white_socks, pink_nails, pink_shirt, autumn_leaves, autumn, pink_sweater,
# of Detected SEGS: 1
model_path is D:\AI\ComfyUI\custom_nodes\ComfyUI-tbox\..\..\models\annotator\TheMistoAI/MistoLine\Anyline\MTEED.pth
model_path is D:\AI\ComfyUI\custom_nodes\ComfyUI-tbox\..\..\models\annotator\yzd-v/DWPose\yolox_l.onnx
model_path is D:\AI\ComfyUI\custom_nodes\ComfyUI-tbox\..\..\models\annotator\hr16/DWPose-TorchScript-BatchSize5\dw-ll_ucoco_384_bs5.torchscript.pt
DWPose: Using yolox_l.onnx for bbox detection and dw-ll_ucoco_384_bs5.torchscript.pt for pose estimation
DWPose: Caching ONNXRuntime session yolox_l.onnx...
DWPose: Caching TorchScript module dw-ll_ucoco_384_bs5.torchscript.pt on ...
DWPose: Bbox 374.71ms
DWPose: Pose 321.67ms on 1 people
model_path is D:\AI\ComfyUI\custom_nodes\ComfyUI-tbox\..\..\models\annotator\LiheYoung/Depth-Anything\checkpoints\depth_anything_vitl14.pth
using MLP layer as FFN
Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
Detailer: force inpaint
Detailer: segment upscale for ((718, 1535)) | crop region (1024, 1536) x 1.0 -> (1024, 1536)
Requested to load AutoencodingEngine
Loading 1 new model
loaded completely 0.0 159.87335777282715 True
Requested to load ControlNet
Requested to load ControlNet
Requested to load ControlNet
Requested to load BaseModel
Loading 4 new models
loaded completely 0.0 689.0852355957031 True
loaded completely 0.0 689.0852355957031 True
loaded completely 0.0 689.0852355957031 True
loaded completely 0.0 1639.406135559082 True
0%| | 0/25 [00:00<?, ?it/s]
!!! Exception during processing !!! Given groups=1, weight of size [320, 4, 3, 3], expected input[2, 16, 192, 128] to have 4 channels, but got 16 channels instead
Traceback (most recent call last):
File "D:\AI\ComfyUI\execution.py", line 323, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\execution.py", line 198, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\execution.py", line 169, in _map_node_over_list
process_inputs(input_dict, i)
File "D:\AI\ComfyUI\execution.py", line 158, in process_inputs
results.append(getattr(obj, func)(**inputs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\impact_pack.py", line 378, in doit
DetailerForEach.do_detail(image, segs, model, clip, vae, guide_size, guide_size_for, max_size, seed, steps,
File "D:\AI\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\impact_pack.py", line 323, in do_detail
enhanced_image, cnet_pils = core.enhance_detail(cropped_image, model, clip, vae, guide_size, guide_size_for_bbox, max_size,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\core.py", line 351, in enhance_detail
refined_latent = impact_sampling.ksampler_wrapper(model2, seed2, steps2, cfg2, sampler_name2, scheduler2, positive2, negative2,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\impact_sampling.py", line 241, in ksampler_wrapper
refined_latent = separated_sample(model, True, seed, advanced_steps, cfg, sampler_name, scheduler,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\impact_sampling.py", line 214, in separated_sample
res = sample_with_custom_noise(model, add_noise, seed, cfg, positive, negative, impact_sampler, sigmas, latent_image, noise=noise, callback=callback)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\custom_nodes\ComfyUI-Impact-Pack\modules\impact\impact_sampling.py", line 158, in sample_with_custom_noise
samples = comfy.sample.sample_custom(model, noise, cfg, sampler, sigmas, positive, negative, latent_image,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\sample.py", line 48, in sample_custom
samples = comfy.samplers.sample(model, noise, positive, negative, cfg, model.load_device, sampler, sigmas, model_options=model.model_options, latent_image=latent_image, denoise_mask=noise_mask, callback=callback, disable_pbar=disable_pbar, seed=seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\samplers.py", line 753, in sample
return cfg_guider.sample(noise, latent_image, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\samplers.py", line 740, in sample
output = self.inner_sample(noise, latent_image, device, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\samplers.py", line 719, in inner_sample
samples = sampler.sample(self, sigmas, extra_args, callback, noise, latent_image, denoise_mask, disable_pbar)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\samplers.py", line 624, in sample
samples = self.sampler_function(model_k, noise, sigmas, extra_args=extra_args, callback=k_callback, disable=disable_pbar, **self.extra_options)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\.venv\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\k_diffusion\sampling.py", line 689, in sample_dpmpp_2m
denoised = model(x, sigmas[i] * s_in, **extra_args)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\samplers.py", line 299, in __call__
out = self.inner_model(x, sigma, model_options=model_options, seed=seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\samplers.py", line 706, in __call__
return self.predict_noise(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\samplers.py", line 709, in predict_noise
return sampling_function(self.inner_model, x, timestep, self.conds.get("negative", None), self.conds.get("positive", None), self.cfg, model_options=model_options, seed=seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\samplers.py", line 279, in sampling_function
out = calc_cond_batch(model, conds, x, timestep, model_options)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\samplers.py", line 202, in calc_cond_batch
c['control'] = control.get_control(input_x, timestep_, c, len(cond_or_uncond))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\controlnet.py", line 203, in get_control
control_prev = self.previous_controlnet.get_control(x_noisy, t, cond, batched_number)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\controlnet.py", line 203, in get_control
control_prev = self.previous_controlnet.get_control(x_noisy, t, cond, batched_number)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\controlnet.py", line 255, in get_control
control = self.control_model(x=x_noisy.to(dtype), hint=self.cond_hint, timesteps=timestep.to(dtype), context=context.to(dtype), **extra)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\.venv\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\.venv\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\cldm\cldm.py", line 426, in forward
h = module(h, emb, context)
^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\.venv\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\.venv\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\ldm\modules\diffusionmodules\openaimodel.py", line 60, in forward
return forward_timestep_embed(self, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\ldm\modules\diffusionmodules\openaimodel.py", line 50, in forward_timestep_embed
x = layer(x)
^^^^^^^^
File "D:\AI\ComfyUI\.venv\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\.venv\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\comfy\ops.py", line 98, in forward
return super().forward(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\.venv\Lib\site-packages\torch\nn\modules\conv.py", line 554, in forward
return self._conv_forward(input, self.weight, self.bias)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\AI\ComfyUI\.venv\Lib\site-packages\torch\nn\modules\conv.py", line 549, in _conv_forward
return F.conv2d(
^^^^^^^^^
RuntimeError: Given groups=1, weight of size [320, 4, 3, 3], expected input[2, 16, 192, 128] to have 4 channels, but got 16 channels instead
Prompt executed in 33.16 seconds
Other
(.venv) PS D:\AI\ComfyUI> python main.py [START] Security scan [DONE] Security scan
ComfyUI-Manager: installing dependencies done.
** ComfyUI startup time: 2024-11-16 20:24:17.560362 ** Platform: Windows ** Python version: 3.12.7 | packaged by Anaconda, Inc. | (main, Oct 4 2024, 13:17:27) [MSC v.1929 64 bit (AMD64)] ** Python executable: D:\AI\ComfyUI.venv\Scripts\python.exe ** ComfyUI Path: D:\AI\ComfyUI ** Log path: D:\AI\ComfyUI\comfyui.log
Prestartup times for custom nodes: 0.0 seconds: D:\AI\ComfyUI\custom_nodes\rgthree-comfy 0.0 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-Easy-Use 5.1 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-Manager
Total VRAM 8188 MB, total RAM 65416 MB pytorch version: 2.5.1+cu124 Set vram state to: NORMAL_VRAM Device: cuda:0 NVIDIA GeForce RTX 4060 : cudaMallocAsync Using pytorch cross attention [Prompt Server] web root: D:\AI\ComfyUI\web [ComfyUI-Easy-Use] server: v1.2.5 Loaded [ComfyUI-Easy-Use] web root: D:\AI\ComfyUI\custom_nodes\ComfyUI-Easy-Use\web_version/v2 Loaded Total VRAM 8188 MB, total RAM 65416 MB pytorch version: 2.5.1+cu124 Set vram state to: NORMAL_VRAM Device: cuda:0 NVIDIA GeForce RTX 4060 : cudaMallocAsync
Loading: ComfyUI-Impact-Pack (V7.11.3)
Loading: ComfyUI-Impact-Pack (Subpack: V0.8)
[Impact Pack] Wildcards loading done. theUpsiders Logic Nodes: Loaded
Loading: ComfyUI-Manager (V2.51.9)
ComfyUI Revision: 2832 [4ac401af] | Released on '2024-11-15'
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json [ReActor] - STATUS - Running v0.5.1-b2 in ComfyUI Torch version: 2.5.1+cu124 no module 'xformers'. Processing without... no module 'xformers'. Processing without... here: D:\AI\ComfyUI\custom_nodes\ComfyUI-tbox Using ckpts path: D:\AI\ComfyUI\custom_nodes\ComfyUI-tbox....\models\annotator Using symlinks: False Using ort providers: ['CUDAExecutionProvider', 'DirectMLExecutionProvider', 'OpenVINOExecutionProvider', 'ROCMExecutionProvider', 'CPUExecutionProvider'] DWPose: Onnxruntime with acceleration providers detected (pysssss:WD14Tagger) [DEBUG] Available ORT providers: TensorrtExecutionProvider, CUDAExecutionProvider, CPUExecutionProvider (pysssss:WD14Tagger) [DEBUG] Using ORT providers: CUDAExecutionProvider, CPUExecutionProvider
Comfyroll Studio v1.76 : 175 Nodes Loaded
** For changes, please see patch notes at https://github.com/Suzie1/ComfyUI_Comfyroll_CustomNodes/blob/main/Patch_Notes.md ** For help, please see the wiki at https://github.com/Suzie1/ComfyUI_Comfyroll_CustomNodes/wiki
[comfyui_controlnet_aux] | INFO -> Using ckpts path: D:\AI\ComfyUI\custom_nodes\comfyui_controlnet_aux\ckpts [comfyui_controlnet_aux] | INFO -> Using symlinks: False [comfyui_controlnet_aux] | INFO -> Using ort providers: ['CUDAExecutionProvider', 'DirectMLExecutionProvider', 'OpenVINOExecutionProvider', 'ROCMExecutionProvider', 'CPUExecutionProvider', 'CoreMLExecutionProvider'] [comfy_mtb] | INFO -> loaded 86 nodes successfuly [comfy_mtb] | INFO -> Some nodes (2) could not be loaded. This can be ignored, but go to http://127.0.0.1:8188/mtb if you want more information. Efficiency Nodes: Attempting to add Control Net options to the 'HiRes-Fix Script' Node (comfyui_controlnet_aux add-on)...Success!
[rgthree-comfy] Loaded 42 epic nodes. 🎉
WAS Node Suite: BlenderNeko's Advanced CLIP Text Encode found, attempting to enable CLIPTextEncode support.
WAS Node Suite: CLIPTextEncode (BlenderNeko Advanced + NSP) node enabled under WAS Suite/Conditioning menu.
WAS Node Suite: OpenCV Python FFMPEG support is enabled
WAS Node Suite: ffmpeg_bin_path is set to: D:/AI/wb/ffmpeg/bin/ffmpeg.exe
WAS Node Suite: Finished. Loaded 219 nodes successfully.
"Art is the universal language that transcends boundaries and speaks to all." - Unknown
Import times for custom nodes: 0.0 seconds: D:\AI\ComfyUI\custom_nodes\websocket_image_save.py 0.0 seconds: D:\AI\ComfyUI\custom_nodes\AIGODLIKE-COMFYUI-TRANSLATION 0.0 seconds: D:\AI\ComfyUI\custom_nodes\Batch-Condition-ComfyUI 0.0 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-Logic 0.0 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI_ADV_CLIP_emb 0.0 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-WD14-Tagger 0.0 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-IC-Light-Native 0.0 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI_IPAdapter_plus 0.0 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-Custom-Scripts 0.0 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI_HF_Servelress_Inference 0.0 seconds: D:\AI\ComfyUI\custom_nodes\comfyui-various 0.0 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI_essentials 0.0 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-KJNodes 0.0 seconds: D:\AI\ComfyUI\custom_nodes\rgthree-comfy 0.1 seconds: D:\AI\ComfyUI\custom_nodes\efficiency-nodes-comfyui 0.1 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-IC-Light 0.1 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI_Comfyroll_CustomNodes 0.1 seconds: D:\AI\ComfyUI\custom_nodes\comfyui_controlnet_aux 0.2 seconds: D:\AI\ComfyUI\custom_nodes\comfyui_segment_anything 0.2 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-Impact-Pack 0.2 seconds: D:\AI\ComfyUI\custom_nodes\Comfyui-ergouzi-Nodes 0.9 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-BrushNet 0.9 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-Easy-Use 0.9 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-Manager 1.2 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-tbox 1.3 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-AdvancedLivePortrait 1.3 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI_LayerStyle 2.2 seconds: D:\AI\ComfyUI\custom_nodes\comfyui-reactor-node 3.2 seconds: D:\AI\ComfyUI\custom_nodes\ComfyUI-SUPIR 3.6 seconds: D:\AI\ComfyUI\custom_nodes\comfy_mtb 5.6 seconds: D:\AI\ComfyUI\custom_nodes\was-node-suite-comfyui
Starting server
To see the GUI go to: http://127.0.0.1:8188
You probably used a model (maybe a ControlNet) that is not compatible with the Checkpoint. You need to check if all models are of the same type (for example all SD1.5)
您可能使用了与 Checkpoint 不兼容的模型(可能是 ControlNet)。 您需要检查所有模型是否属于同一类型(例如所有 SD1.5)
Thank you very much! After switching the VAE model, everything is working fine now.
您可能使用了与Checkpoint不兼容的模型(可能是ControlNet)。您需要检查所有模型是否属于相同类型(例如所有SD1.5)
非常感谢!切换 VAE 模型后,现在一切都正常了。
怎么弄的哥,用的flux的vae没错吧,还有别的吗
"model_path is D:\AI\ComfyUI\custom_nodes\ComfyUI-tbox....\models\annotator\TheMistoAI/MistoLine\Anyline\MTEED.pth" 我想知道这个“....\”中间具体是什么内容!我离线下载了模型,不知道往哪放!就差中间省略的路径!
You probably used a model (maybe a ControlNet) that is not compatible with the Checkpoint. You need to check if all models are of the same type (for example all SD1.5)
i dont see VAE in this workflow and I'm getting this error. here is the workflow: https://www.youtube.com/watch?v=9L4y-SIw6RU