[dask] PBSCluster - nothing is executed - pipeline freezes #4237
Unanswered
erwanbodere
asked this question in
Q&A
Replies: 2 comments 3 replies
-
Hi,
# Allow interrupts while waiting for the results from Dask
for future, result in iterate_with_context(raise_execution_interrupts, futures):
for step_event in result:
check.inst(step_event, DagsterEvent)
yield step_event
# Executes the next() function within an instance of the supplied context manager class
# (leaving the context before yielding each result)
def iterate_with_context(context, iterator):
while True:
# Allow interrupts during user code so that we can terminate slow/hanging steps
with context():
try:
next_output = next(iterator)
except StopIteration:
return
yield next_output Nothing is executed at Here are the steps to reproduce :
from dagster import ModeDefinition, default_executors, fs_io_manager, pipeline
from dagster_dask import dask_executor
from dagster_shell import create_shell_command_solid
@pipeline(
mode_defs=[
ModeDefinition(
resource_defs={"io_manager": fs_io_manager},
executor_defs=default_executors + [dask_executor],
)
]
)
def hello_pipeline():
cmd_solid = create_shell_command_solid('echo "hello, world!"', name="hello_world")
cmd_solid()
execution:
dask:
config:
cluster:
pbs:
memory: '24GB'
cores: 1
loggers:
console:
config:
log_level: DEBUG |
Beta Was this translation helpful? Give feedback.
0 replies
-
Hi, Maybe dagster is not working with dask-jobqueue ? The LocalCluster is using distributed.LocalCluster and not dask_jobqueue.local.LocalCluster. Any idea ? Rregards |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
I want to execute a pipeline with dagster on an HPC cluster using PBS Pro.
The configuration for PBS can be found here : https://github.com/dask/dask-jobqueue/blob/main/dask_jobqueue/jobqueue.yaml
I can execute a process with Dask on our HPC cluster but i don't achieve to execute a pipeline with dasgter. Dasgter freezes at "PIPELINE START - Started execution of pipeline", without any errors.
Here is my yaml file :
Is ti possible to have more logs about Dask, for instance with
cluster.job_script()
in order to understand what's happen ?See : https://jobqueue.dask.org/en/latest/debug.html
Regards
Beta Was this translation helpful? Give feedback.
All reactions