Skip to content

Validate circular reference for yaml (#33208) #12811

Validate circular reference for yaml (#33208)

Validate circular reference for yaml (#33208) #12811

GitHub Actions / Test Results failed Dec 18, 2024 in 0s

1 fail, 22 skipped, 2 pass in 6m 42s

25 tests    2 ✅  6m 42s ⏱️
 1 suites  22 💤
 1 files     1 ❌

Results for commit e68a79c.

Annotations

Check warning on line 0 in apache_beam.transforms.managed_iceberg_it_test.ManagedIcebergIT

See this annotation in the file changed.

@github-actions github-actions / Test Results

test_write_read_pipeline (apache_beam.transforms.managed_iceberg_it_test.ManagedIcebergIT) failed

sdks/python/pytest_ioCrossLanguage.xml [took 5m 55s]
Raw output
apache_beam.runners.dataflow.dataflow_runner.DataflowRuntimeException: Dataflow pipeline failed. State: FAILED, Error:
None
self = <apache_beam.transforms.managed_iceberg_it_test.ManagedIcebergIT testMethod=test_write_read_pipeline>

    def test_write_read_pipeline(self):
      iceberg_config = {
          "table": "test_iceberg_write_read.test_" + str(int(time.time())),
          "catalog_name": "default",
          "catalog_properties": {
              "type": "hadoop",
              "warehouse": self.WAREHOUSE,
          }
      }
    
      rows = [self._create_row(i) for i in range(100)]
      expected_dicts = [row.as_dict() for row in rows]
    
      with beam.Pipeline(argv=self.args) as write_pipeline:
>       _ = (
            write_pipeline
            | beam.Create(rows)
            | beam.managed.Write(beam.managed.ICEBERG, config=iceberg_config))

apache_beam/transforms/managed_iceberg_it_test.py:68: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:620: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:594: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/dataflow/test_dataflow_runner.py:66: in run_pipeline
    self.result.wait_until_finish(duration=wait_duration)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

self = <DataflowPipelineResult <Job
 clientRequestId: '20241218055550270621-4304'
 createTime: '2024-12-18T05:55:55.698506Z'
...024-12-18T05:55:55.698506Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)> at 0x7c86f1f60640>
duration = None

    def wait_until_finish(self, duration=None):
      if not self.is_in_terminal_state():
        if not self.has_job:
          raise IOError('Failed to get the Dataflow job id.')
        consoleUrl = (
            "Console URL: https://console.cloud.google.com/"
            f"dataflow/jobs/<RegionId>/{self.job_id()}"
            "?project=<ProjectId>")
        thread = threading.Thread(
            target=DataflowRunner.poll_for_job_completion,
            args=(self._runner, self, duration))
    
        # Mark the thread as a daemon thread so a keyboard interrupt on the main
        # thread will terminate everything. This is also the reason we will not
        # use thread.join() to wait for the polling thread.
        thread.daemon = True
        thread.start()
        while thread.is_alive():
          time.sleep(5.0)
    
        # TODO: Merge the termination code in poll_for_job_completion and
        # is_in_terminal_state.
        terminated = self.is_in_terminal_state()
        assert duration or terminated, (
            'Job did not reach to a terminal state after waiting indefinitely. '
            '{}'.format(consoleUrl))
    
        if terminated and self.state != PipelineState.DONE:
          # TODO(BEAM-1290): Consider converting this to an error log based on
          # theresolution of the issue.
          _LOGGER.error(consoleUrl)
          raise DataflowRuntimeException(
              'Dataflow pipeline failed. State: %s, Error:\n%s' %
              (self.state, getattr(self._runner, 'last_error_msg', None)),
              self)
      elif PipelineState.is_terminal(
          self.state) and self.state == PipelineState.FAILED and self._runner:
>       raise DataflowRuntimeException(
            'Dataflow pipeline failed. State: %s, Error:\n%s' %
            (self.state, getattr(self._runner, 'last_error_msg', None)),
            self)
E       apache_beam.runners.dataflow.dataflow_runner.DataflowRuntimeException: Dataflow pipeline failed. State: FAILED, Error:
E       None

apache_beam/runners/dataflow/dataflow_runner.py:813: DataflowRuntimeException

Check notice on line 0 in .github

See this annotation in the file changed.

@github-actions github-actions / Test Results

22 skipped tests found

There are 22 skipped tests, see "Raw output" for the full list of skipped tests.
Raw output
apache_beam.examples.ml_transform.ml_transform_it_test
apache_beam.examples.snippets.transforms.elementwise.mltransform_test
apache_beam.examples.snippets.transforms.elementwise.runinference_test
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_null_key
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_populated_key
apache_beam.ml.inference.huggingface_inference_it_test
apache_beam.ml.inference.huggingface_inference_test
apache_beam.ml.inference.onnx_inference_test
apache_beam.ml.inference.pytorch_inference_test
apache_beam.ml.inference.tensorflow_inference_test
apache_beam.ml.inference.tensorrt_inference_test
apache_beam.ml.inference.vertex_ai_inference_it_test
apache_beam.ml.inference.xgboost_inference_test
apache_beam.ml.transforms.handlers_test
apache_beam.ml.transforms.tft_test
apache_beam.runners.dask.dask_runner_test
apache_beam.testing.analyzers.perf_analysis_test
apache_beam.testing.benchmarks.cloudml.cloudml_benchmark_test
apache_beam.transforms.enrichment_handlers.feast_feature_store_it_test
apache_beam.transforms.enrichment_handlers.feast_feature_store_test
apache_beam.typehints.pytorch_type_compatibility_test
apache_beam.yaml.yaml_ml_test

Check notice on line 0 in .github

See this annotation in the file changed.

@github-actions github-actions / Test Results

25 tests found

There are 25 tests, see "Raw output" for the full list of tests.
Raw output
apache_beam.examples.ml_transform.ml_transform_it_test
apache_beam.examples.snippets.transforms.elementwise.mltransform_test
apache_beam.examples.snippets.transforms.elementwise.runinference_test
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_null_key
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_populated_key
apache_beam.ml.inference.huggingface_inference_it_test
apache_beam.ml.inference.huggingface_inference_test
apache_beam.ml.inference.onnx_inference_test
apache_beam.ml.inference.pytorch_inference_test
apache_beam.ml.inference.tensorflow_inference_test
apache_beam.ml.inference.tensorrt_inference_test
apache_beam.ml.inference.vertex_ai_inference_it_test
apache_beam.ml.inference.xgboost_inference_test
apache_beam.ml.transforms.handlers_test
apache_beam.ml.transforms.tft_test
apache_beam.runners.dask.dask_runner_test
apache_beam.testing.analyzers.perf_analysis_test
apache_beam.testing.benchmarks.cloudml.cloudml_benchmark_test
apache_beam.transforms.enrichment_handlers.feast_feature_store_it_test
apache_beam.transforms.enrichment_handlers.feast_feature_store_test
apache_beam.transforms.external_transform_provider_it_test.ExternalTransformProviderIT ‑ test_generate_sequence_signature_and_doc
apache_beam.transforms.external_transform_provider_it_test.ExternalTransformProviderIT ‑ test_run_generate_sequence
apache_beam.transforms.managed_iceberg_it_test.ManagedIcebergIT ‑ test_write_read_pipeline
apache_beam.typehints.pytorch_type_compatibility_test
apache_beam.yaml.yaml_ml_test