We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Traceback (most recent call last): File "D:\ComfyUI-aki-v1.2\execution.py", line 323, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "D:\ComfyUI-aki-v1.2\execution.py", line 198, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "D:\ComfyUI-aki-v1.2\execution.py", line 169, in _map_node_over_list process_inputs(input_dict, i) File "D:\ComfyUI-aki-v1.2\execution.py", line 158, in process_inputs results.append(getattr(obj, func)(**inputs)) File "D:\ComfyUI-aki-v1.2\custom_nodes\BizyAir\src\bizyair\nodes_base.py", line 70, in new_func return org_func(self, **kwargs) File "D:\ComfyUI-aki-v1.2\custom_nodes\BizyAir\src\bizyair\nodes_base.py", line 129, in default_function return self._process_all_send_request_types(node_ios) File "D:\ComfyUI-aki-v1.2\custom_nodes\BizyAir\src\bizyair\nodes_base.py", line 156, in _process_all_send_request_types return node_ios[0].send_request() File "D:\ComfyUI-aki-v1.2\custom_nodes\BizyAir\src\bizyair\nodes_io.py", line 88, in send_request out = invoker.prompt_server.execute( File "D:\ComfyUI-aki-v1.2\custom_nodes\BizyAir\src\bizyair\commands\servers\prompt_server.py", line 45, in execute raise RuntimeError( RuntimeError: Unexpected error accessing result["data"]["payload"]. Result: {'type': 'error', 'message': 'Error processing task 8152369a-68bc-419e-94b3-7a193771b551: Exceeded limit for instance ModelSamplingFlux count <= 0.', 'details': 'Traceback (most recent call last):\n File "/workspace/comfybridge/comfybridge/bizyair/schedulers.py", line 287, in process_request\n result = await self._execute_task(\n File "/workspace/comfybridge/comfybridge/bizyair/schedulers.py", line 341, in _execute_task\n return await loop.run_in_executor(self.pool, executor.execute)\n File "/opt/conda/lib/python3.10/concurrent/futures/thread.py", line 58, in run\n result = self.fn(*self.args, **self.kwargs)\n File "/workspace/comfybridge/comfybridge/bizyair/schedulers.py", line 80, in execute\n return self.execute_func(*self.args, **self.kwargs)\n File "/workspace/comfybridge/comfybridge/plugins/comfy_pipeline/utils/cuda_memory_utils.py", line 22, in wrapper\n result = func(*args, **kwargs)\n File "/opt/conda/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context\n return func(*args, **kwargs)\n File "/workspace/comfybridge/comfybridge/plugins/comfy_pipeline/server.py", line 124, in run\n user_params: ServerFormat = convert_to_server_format(\n File "/workspace/comfybridge/comfybridge/plugins/comfy_pipeline/transpiler/transpiler.py", line 81, in convert_to_server_format\n prompt, node_id_registry = allocator.generate_workflow_with_allocated_ids(\n File "/workspace/comfybridge/comfybridge/plugins/comfy_pipeline/transpiler/node_id_allocator.py", line 105, in generate_workflow_with_allocated_ids\n raise RuntimeError(\nRuntimeError: Exceeded limit for instance ModelSamplingFlux count <= 0\n', 'extra_info': None, 'data_status': 'COMPLETED', 'start_time': 1731586702.1202548, 'end_time': 1731586702.1833594, 'task_id': '8152369a-68bc-419e-94b3-7a193771b551', 'upload_to_s3': None}
The text was updated successfully, but these errors were encountered:
No branches or pull requests
Traceback (most recent call last):
File "D:\ComfyUI-aki-v1.2\execution.py", line 323, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
File "D:\ComfyUI-aki-v1.2\execution.py", line 198, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
File "D:\ComfyUI-aki-v1.2\execution.py", line 169, in _map_node_over_list
process_inputs(input_dict, i)
File "D:\ComfyUI-aki-v1.2\execution.py", line 158, in process_inputs
results.append(getattr(obj, func)(**inputs))
File "D:\ComfyUI-aki-v1.2\custom_nodes\BizyAir\src\bizyair\nodes_base.py", line 70, in new_func
return org_func(self, **kwargs)
File "D:\ComfyUI-aki-v1.2\custom_nodes\BizyAir\src\bizyair\nodes_base.py", line 129, in default_function
return self._process_all_send_request_types(node_ios)
File "D:\ComfyUI-aki-v1.2\custom_nodes\BizyAir\src\bizyair\nodes_base.py", line 156, in _process_all_send_request_types
return node_ios[0].send_request()
File "D:\ComfyUI-aki-v1.2\custom_nodes\BizyAir\src\bizyair\nodes_io.py", line 88, in send_request
out = invoker.prompt_server.execute(
File "D:\ComfyUI-aki-v1.2\custom_nodes\BizyAir\src\bizyair\commands\servers\prompt_server.py", line 45, in execute
raise RuntimeError(
RuntimeError: Unexpected error accessing result["data"]["payload"]. Result: {'type': 'error', 'message': 'Error processing task 8152369a-68bc-419e-94b3-7a193771b551: Exceeded limit for instance ModelSamplingFlux count <= 0.', 'details': 'Traceback (most recent call last):\n File "/workspace/comfybridge/comfybridge/bizyair/schedulers.py", line 287, in process_request\n result = await self._execute_task(\n File "/workspace/comfybridge/comfybridge/bizyair/schedulers.py", line 341, in _execute_task\n return await loop.run_in_executor(self.pool, executor.execute)\n File "/opt/conda/lib/python3.10/concurrent/futures/thread.py", line 58, in run\n result = self.fn(*self.args, **self.kwargs)\n File "/workspace/comfybridge/comfybridge/bizyair/schedulers.py", line 80, in execute\n return self.execute_func(*self.args, **self.kwargs)\n File "/workspace/comfybridge/comfybridge/plugins/comfy_pipeline/utils/cuda_memory_utils.py", line 22, in wrapper\n result = func(*args, **kwargs)\n File "/opt/conda/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context\n return func(*args, **kwargs)\n File "/workspace/comfybridge/comfybridge/plugins/comfy_pipeline/server.py", line 124, in run\n user_params: ServerFormat = convert_to_server_format(\n File "/workspace/comfybridge/comfybridge/plugins/comfy_pipeline/transpiler/transpiler.py", line 81, in convert_to_server_format\n prompt, node_id_registry = allocator.generate_workflow_with_allocated_ids(\n File "/workspace/comfybridge/comfybridge/plugins/comfy_pipeline/transpiler/node_id_allocator.py", line 105, in generate_workflow_with_allocated_ids\n raise RuntimeError(\nRuntimeError: Exceeded limit for instance ModelSamplingFlux count <= 0\n', 'extra_info': None, 'data_status': 'COMPLETED', 'start_time': 1731586702.1202548, 'end_time': 1731586702.1833594, 'task_id': '8152369a-68bc-419e-94b3-7a193771b551', 'upload_to_s3': None}
The text was updated successfully, but these errors were encountered: