-
Notifications
You must be signed in to change notification settings - Fork 30
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error running DragRun node #3
Comments
+1 |
作者好像一直在开发新的 老问题都不怎么修 |
这个报错原因是xformers版本和cuda版本不匹配 cuda 11.8 versionpip3 install -U xformers --index-url https://download.pytorch.org/whl/cu118 cuda 12.1 versionpip3 install -U xformers --index-url https://download.pytorch.org/whl/cu121https://github.com/facebookresearch/xformers#installing-xformers |
Translation? |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I downloaded the model and placed it in the right location, but I still get this error:
'''' 0%| | 0/1 [00:00<?, ?it/s]
ERROR:root:!!! Exception during processing !!!
ERROR:root:Traceback (most recent call last):
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 153, in recursive_execute
output_data, output_ui = get_output_data(obj, input_data_all)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 83, in get_output_data
return_values = map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 76, in map_node_over_list
results.append(getattr(obj, func)(**slice_dict(input_data_all, i)))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-DragNUWA\nodes.py", line 251, in run_inference
return model.run(image_pil, tracking_points, inference_batch_size, motion_bucket_id)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-DragNUWA\nodes.py", line 185, in run
outputs = self.forward_sample(
^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-DragNUWA\nodes.py", line 108, in forward_sample
c, uc = self.drag_nuwa_net.conditioner.get_unconditional_conditioning(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-DragNUWA\dragnuwa\svd\modules\encoders\modules.py", line 180, in get_unconditional_conditioning
c = self(batch_c, force_cond_zero_embeddings)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-DragNUWA\dragnuwa\svd\modules\encoders\modules.py", line 133, in forward
emb_out = embedder(batch[embedder.input_key])
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-DragNUWA\dragnuwa\svd\modules\encoders\modules.py", line 1013, in forward
out = self.encoder.encode(vid[n * n_samples : (n + 1) * n_samples])
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-DragNUWA\dragnuwa\svd\models\autoencoder.py", line 472, in encode
z = self.encoder(x)
^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1518, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1527, in _call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-DragNUWA\dragnuwa\svd\modules\diffusionmodules\model.py", line 594, in forward
h = self.mid.attn_1(h)
^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1518, in wrapped_call_impl
return self.call_impl(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1527, in call_impl
return forward_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-DragNUWA\dragnuwa\svd\modules\diffusionmodules\model.py", line 263, in forward
h = self.attention(h)
^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-DragNUWA\dragnuwa\svd\modules\diffusionmodules\model.py", line 249, in attention
out = xformers.ops.memory_efficient_attention(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\xformers\ops\fmha_init.py", line 223, in memory_efficient_attention
return memory_efficient_attention(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\xformers\ops\fmha_init.py", line 321, in _memory_efficient_attention
return memory_efficient_attention_forward(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\xformers\ops\fmha_init.py", line 337, in _memory_efficient_attention_forward
op = _dispatch_fw(inp, False)
^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\xformers\ops\fmha\dispatch.py", line 120, in _dispatch_fw
return _run_priority_list(
^^^^^^^^^^^^^^^^^^^
File "E:\COMFYUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\xformers\ops\fmha\dispatch.py", line 63, in _run_priority_list
raise NotImplementedError(msg)
NotImplementedError: No operator found for
memory_efficient_attention_forward
with inputs:query : shape=(1, 2880, 1, 512) (torch.float32)
key : shape=(1, 2880, 1, 512) (torch.float32)
value : shape=(1, 2880, 1, 512) (torch.float32)
attn_bias : <class 'NoneType'>
p : 0.0
decoderF
is not supported because:max(query.shape[-1] != value.shape[-1]) > 128
xFormers wasn't build with CUDA support
attn_bias type is <class 'NoneType'>
operator wasn't built - see
python -m xformers.info
for more info[email protected]
is not supported because:max(query.shape[-1] != value.shape[-1]) > 256
xFormers wasn't build with CUDA support
dtype=torch.float32 (supported: {torch.float16, torch.bfloat16})
operator wasn't built - see
python -m xformers.info
for more infotritonflashattF
is not supported because:max(query.shape[-1] != value.shape[-1]) > 128
xFormers wasn't build with CUDA support
dtype=torch.float32 (supported: {torch.float16, torch.bfloat16})
operator wasn't built - see
python -m xformers.info
for more infotriton is not available
cutlassF
is not supported because:xFormers wasn't build with CUDA support
operator wasn't built - see
python -m xformers.info
for more infosmallkF
is not supported because:max(query.shape[-1] != value.shape[-1]) > 32
xFormers wasn't build with CUDA support
operator wasn't built - see
python -m xformers.info
for more infounsupported embed per head: 512
Prompt execu'''
The text was updated successfully, but these errors were encountered: