Skip to content

[DO NOT MERGE] Run all PostCommit and PreCommit Tests against Release Branch #4106

[DO NOT MERGE] Run all PostCommit and PreCommit Tests against Release Branch

[DO NOT MERGE] Run all PostCommit and PreCommit Tests against Release Branch #4106

GitHub Actions / Test Results failed Dec 19, 2024 in 0s

1 fail, 19 skipped, 2 pass in 48m 59s

22 tests    2 ✅  48m 59s ⏱️
 1 suites  19 💤
 1 files     1 ❌

Results for commit c1ba1bd.

Annotations

Check warning on line 0 in apache_beam.examples.wordcount_it_test.WordCountIT

See this annotation in the file changed.

@github-actions github-actions / Test Results

test_wordcount_it_with_prebuilt_sdk_container_local_docker (apache_beam.examples.wordcount_it_test.WordCountIT) failed

sdks/python/pytest-beam_python3.12_sdk.xml [took 33m 1s]
Raw output
Failed: Timeout >1800.0s
self = <apache_beam.examples.wordcount_it_test.WordCountIT testMethod=test_wordcount_it_with_prebuilt_sdk_container_local_docker>

    @pytest.mark.it_validatescontainer
    def test_wordcount_it_with_prebuilt_sdk_container_local_docker(self):
>     self._run_wordcount_it(
          wordcount.run,
          experiment='beam_fn_api',
          prebuild_sdk_container_engine='local_docker')

apache_beam/examples/wordcount_it_test.py:95: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/examples/wordcount_it_test.py:150: in _run_wordcount_it
    run_wordcount(
apache_beam/examples/wordcount.py:109: in run
    result = pipeline.run()
apache_beam/pipeline.py:570: in run
    self._options).run(False)
apache_beam/pipeline.py:594: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/dataflow/test_dataflow_runner.py:66: in run_pipeline
    self.result.wait_until_finish(duration=wait_duration)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <DataflowPipelineResult <Job
 clientRequestId: '20241219150012135050-7520'
 createTime: '2024-12-19T15:00:13.080921Z'
...024-12-19T15:00:13.080921Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)> at 0x7b3264e42a20>
duration = None

    def wait_until_finish(self, duration=None):
      if not self.is_in_terminal_state():
        if not self.has_job:
          raise IOError('Failed to get the Dataflow job id.')
        consoleUrl = (
            "Console URL: https://console.cloud.google.com/"
            f"dataflow/jobs/<RegionId>/{self.job_id()}"
            "?project=<ProjectId>")
        thread = threading.Thread(
            target=DataflowRunner.poll_for_job_completion,
            args=(self._runner, self, duration))
    
        # Mark the thread as a daemon thread so a keyboard interrupt on the main
        # thread will terminate everything. This is also the reason we will not
        # use thread.join() to wait for the polling thread.
        thread.daemon = True
        thread.start()
        while thread.is_alive():
>         time.sleep(5.0)
E         Failed: Timeout >1800.0s

apache_beam/runners/dataflow/dataflow_runner.py:794: Failed