Merge pull request #32791 from liferoad/beam-summit-2024 #29920
GitHub Actions / Test Results
failed
Oct 16, 2024 in 0s
1 fail, 22 skipped, 2 pass in 1m 4s
25 tests 2 ✅ 1m 4s ⏱️
1 suites 22 💤
1 files 1 ❌
Results for commit 1487cf5.
Annotations
Check warning on line 0 in apache_beam.transforms.managed_iceberg_it_test.ManagedIcebergIT
github-actions / Test Results
test_write_read_pipeline (apache_beam.transforms.managed_iceberg_it_test.ManagedIcebergIT) failed
sdks/python/pytest_ioCrossLanguage.xml [took 10s]
Raw output
RuntimeError: java.lang.IllegalArgumentException: unable to serialize SchemaCoder<Schema: Fields:
Field{name=tableIdentifierString, description=, type=STRING NOT NULL, options={{}}}
Field{name=serializableDataFile, description=, type=ROW<fileFormat STRING NOT NULL, recordCount INT64 NOT NULL, fileSizeInBytes INT64 NOT NULL, splitOffsets ARRAY<INT64 NOT NULL>, keyMetadata BYTES, partitionPath STRING NOT NULL, partitionSpecId INT32 NOT NULL, columnSizes MAP<INT32 NOT NULL, INT64 NOT NULL>, valueCounts MAP<INT32 NOT NULL, INT64 NOT NULL>, nullValueCounts MAP<INT32 NOT NULL, INT64 NOT NULL>, nanValueCounts MAP<INT32 NOT NULL, INT64 NOT NULL>, upperBounds MAP<INT32 NOT NULL, BYTES NOT NULL>, lowerBounds MAP<INT32 NOT NULL, BYTES NOT NULL>, path STRING NOT NULL> NOT NULL, options={{}}}
Encoding positions:
{tableIdentifierString=0, serializableDataFile=1}
Options:{{}}UUID: f534fce0-5357-451c-8993-8c5fb3b75bca UUID: f534fce0-5357-451c-8993-8c5fb3b75bca delegateCoder: org.apache.beam.sdk.coders.Coder$ByteBuddy$mdNrLYeU@2f2de4a7
at org.apache.beam.sdk.util.SerializableUtils.serializeToByteArray(SerializableUtils.java:59)
at org.apache.beam.sdk.util.construction.CoderTranslation.toCustomCoder(CoderTranslation.java:158)
at org.apache.beam.sdk.util.construction.CoderTranslation.toProto(CoderTranslation.java:118)
at org.apache.beam.sdk.util.construction.SdkComponents.registerCoder(SdkComponents.java:284)
at org.apache.beam.sdk.util.construction.PCollectionTranslation.toProto(PCollectionTranslation.java:35)
at org.apache.beam.sdk.util.construction.SdkComponents.registerPCollection(SdkComponents.java:239)
at org.apache.beam.sdk.util.construction.PTransformTranslation.translateAppliedPTransform(PTransformTranslation.java:610)
at org.apache.beam.sdk.util.construction.ParDoTranslation$ParDoTranslator.translate(ParDoTranslation.java:184)
at org.apache.beam.sdk.util.construction.PTransformTranslation.toProto(PTransformTranslation.java:277)
at org.apache.beam.sdk.util.construction.SdkComponents.registerPTransform(SdkComponents.java:183)
at org.apache.beam.sdk.util.construction.PipelineTranslation$1.visitPrimitiveTransform(PipelineTranslation.java:96)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:593)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:585)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:585)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:585)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:585)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:585)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:585)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.access$500(TransformHierarchy.java:240)
at org.apache.beam.sdk.runners.TransformHierarchy.visit(TransformHierarchy.java:214)
at org.apache.beam.sdk.Pipeline.traverseTopologically(Pipeline.java:477)
at org.apache.beam.sdk.util.construction.PipelineTranslation.toProto(PipelineTranslation.java:68)
at org.apache.beam.sdk.util.construction.PipelineTranslation.toProto(PipelineTranslation.java:59)
at org.apache.beam.sdk.util.construction.PipelineTranslation.toProto(PipelineTranslation.java:52)
at org.apache.beam.sdk.expansion.service.ExpansionService.expand(ExpansionService.java:701)
at org.apache.beam.sdk.expansion.service.ExpansionService.expand(ExpansionService.java:762)
at org.apache.beam.model.expansion.v1.ExpansionServiceGrpc$MethodHandlers.invoke(ExpansionServiceGrpc.java:306)
at org.apache.beam.vendor.grpc.v1p60p1.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:182)
at org.apache.beam.vendor.grpc.v1p60p1.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:351)
at org.apache.beam.vendor.grpc.v1p60p1.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:861)
at org.apache.beam.vendor.grpc.v1p60p1.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at org.apache.beam.vendor.grpc.v1p60p1.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: java.io.NotSerializableException: org.apache.beam.sdk.schemas.utils.ByteBuddyUtils$TransformingMap
at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1184)
at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:348)
at org.apache.beam.sdk.util.SerializableUtils.serializeToByteArray(SerializableUtils.java:55)
... 34 more
java.lang.IllegalArgumentException: unable to serialize SchemaCoder<Schema: Fields:
Field{name=tableIdentifierString, description=, type=STRING NOT NULL, options={{}}}
Field{name=serializableDataFile, description=, type=ROW<fileFormat STRING NOT NULL, recordCount INT64 NOT NULL, fileSizeInBytes INT64 NOT NULL, splitOffsets ARRAY<INT64 NOT NULL>, keyMetadata BYTES, partitionPath STRING NOT NULL, partitionSpecId INT32 NOT NULL, columnSizes MAP<INT32 NOT NULL, INT64 NOT NULL>, valueCounts MAP<INT32 NOT NULL, INT64 NOT NULL>, nullValueCounts MAP<INT32 NOT NULL, INT64 NOT NULL>, nanValueCounts MAP<INT32 NOT NULL, INT64 NOT NULL>, upperBounds MAP<INT32 NOT NULL, BYTES NOT NULL>, lowerBounds MAP<INT32 NOT NULL, BYTES NOT NULL>, path STRING NOT NULL> NOT NULL, options={{}}}
Encoding positions:
{tableIdentifierString=0, serializableDataFile=1}
Options:{{}}UUID: f534fce0-5357-451c-8993-8c5fb3b75bca UUID: f534fce0-5357-451c-8993-8c5fb3b75bca delegateCoder: org.apache.beam.sdk.coders.Coder$ByteBuddy$mdNrLYeU@2f2de4a7
self = <apache_beam.transforms.managed_iceberg_it_test.ManagedIcebergIT testMethod=test_write_read_pipeline>
def test_write_read_pipeline(self):
iceberg_config = {
"table": "test.write_read",
"catalog_name": "default",
"catalog_properties": {
"type": "hadoop",
"warehouse": f"file://{self.warehouse_path}",
}
}
rows = [self._create_row(i) for i in range(100)]
expected_dicts = [row.as_dict() for row in rows]
with beam.Pipeline() as write_pipeline:
_ = (
> write_pipeline
| beam.Create(rows)
| beam.managed.Write(beam.managed.ICEBERG, config=iceberg_config))
apache_beam/transforms/managed_iceberg_it_test.py:73:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pvalue.py:138: in __or__
return self.pipeline.apply(ptransform, self)
apache_beam/pipeline.py:754: in apply
pvalueish_result = self.runner.apply(transform, pvalueish, self._options)
apache_beam/runners/runner.py:191: in apply
return self.apply_PTransform(transform, input, options)
apache_beam/runners/runner.py:195: in apply_PTransform
return transform.expand(input)
apache_beam/transforms/managed.py:156: in expand
return input | SchemaAwareExternalTransform(
apache_beam/pvalue.py:138: in __or__
return self.pipeline.apply(ptransform, self)
apache_beam/pipeline.py:754: in apply
pvalueish_result = self.runner.apply(transform, pvalueish, self._options)
apache_beam/runners/runner.py:191: in apply
return self.apply_PTransform(transform, input, options)
apache_beam/runners/runner.py:195: in apply_PTransform
return transform.expand(input)
apache_beam/transforms/external.py:429: in expand
return pcolls | self._payload_builder.identifier() >> ExternalTransform(
apache_beam/pvalue.py:138: in __or__
return self.pipeline.apply(ptransform, self)
apache_beam/pipeline.py:681: in apply
return self.apply(
apache_beam/pipeline.py:692: in apply
return self.apply(transform, pvalueish)
apache_beam/pipeline.py:754: in apply
pvalueish_result = self.runner.apply(transform, pvalueish, self._options)
apache_beam/runners/runner.py:191: in apply
return self.apply_PTransform(transform, input, options)
apache_beam/runners/runner.py:195: in apply_PTransform
return transform.expand(input)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <ExternalTransform(PTransform) label=[ExternalTransform(beam:expansion:payload:schematransform:v1)] at 0x7aeb9057e840>
pvalueish = <PCollection[Create/Map(decode).None] at 0x7aeb906d2de0>
def expand(self, pvalueish):
# type: (pvalue.PCollection) -> pvalue.PCollection
if isinstance(pvalueish, pvalue.PBegin):
self._inputs = {}
elif isinstance(pvalueish, (list, tuple)):
self._inputs = {str(ix): pvalue for ix, pvalue in enumerate(pvalueish)}
elif isinstance(pvalueish, dict):
self._inputs = pvalueish
else:
self._inputs = {'input': pvalueish}
pipeline = (
next(iter(self._inputs.values())).pipeline
if self._inputs else pvalueish.pipeline)
context = pipeline_context.PipelineContext(
component_id_map=pipeline.component_id_map)
transform_proto = beam_runner_api_pb2.PTransform(
unique_name=pipeline._current_transform().full_label,
spec=beam_runner_api_pb2.FunctionSpec(
urn=self._urn, payload=self._payload))
for tag, pcoll in self._inputs.items():
transform_proto.inputs[tag] = context.pcollections.get_id(pcoll)
# Conversion to/from proto assumes producers.
# TODO: Possibly loosen this.
context.transforms.put_proto(
'%s_%s' % (self._IMPULSE_PREFIX, tag),
beam_runner_api_pb2.PTransform(
unique_name='%s_%s' % (self._IMPULSE_PREFIX, tag),
spec=beam_runner_api_pb2.FunctionSpec(
urn=common_urns.primitives.IMPULSE.urn),
outputs={'out': transform_proto.inputs[tag]}))
output_coders = None
if self._type_hints.output_types:
if self._type_hints.output_types[0]:
output_coders = dict(
(str(k), context.coder_id_from_element_type(v))
for (k, v) in enumerate(self._type_hints.output_types[0]))
elif self._type_hints.output_types[1]:
output_coders = {
k: context.coder_id_from_element_type(v)
for (k, v) in self._type_hints.output_types[1].items()
}
components = context.to_runner_api()
request = beam_expansion_api_pb2.ExpansionRequest(
components=components,
namespace=self._external_namespace,
transform=transform_proto,
output_coder_requests=output_coders,
pipeline_options=pipeline._options.to_runner_api())
expansion_service = _maybe_use_transform_service(
self._expansion_service, pipeline.options)
with ExternalTransform.service(expansion_service) as service:
response = service.Expand(request)
if response.error:
> raise RuntimeError(_sanitize_java_traceback(response.error))
E RuntimeError: java.lang.IllegalArgumentException: unable to serialize SchemaCoder<Schema: Fields:
E Field{name=tableIdentifierString, description=, type=STRING NOT NULL, options={{}}}
E Field{name=serializableDataFile, description=, type=ROW<fileFormat STRING NOT NULL, recordCount INT64 NOT NULL, fileSizeInBytes INT64 NOT NULL, splitOffsets ARRAY<INT64 NOT NULL>, keyMetadata BYTES, partitionPath STRING NOT NULL, partitionSpecId INT32 NOT NULL, columnSizes MAP<INT32 NOT NULL, INT64 NOT NULL>, valueCounts MAP<INT32 NOT NULL, INT64 NOT NULL>, nullValueCounts MAP<INT32 NOT NULL, INT64 NOT NULL>, nanValueCounts MAP<INT32 NOT NULL, INT64 NOT NULL>, upperBounds MAP<INT32 NOT NULL, BYTES NOT NULL>, lowerBounds MAP<INT32 NOT NULL, BYTES NOT NULL>, path STRING NOT NULL> NOT NULL, options={{}}}
E Encoding positions:
E {tableIdentifierString=0, serializableDataFile=1}
E Options:{{}}UUID: f534fce0-5357-451c-8993-8c5fb3b75bca UUID: f534fce0-5357-451c-8993-8c5fb3b75bca delegateCoder: org.apache.beam.sdk.coders.Coder$ByteBuddy$mdNrLYeU@2f2de4a7
E at org.apache.beam.sdk.util.SerializableUtils.serializeToByteArray(SerializableUtils.java:59)
E at org.apache.beam.sdk.util.construction.CoderTranslation.toCustomCoder(CoderTranslation.java:158)
E at org.apache.beam.sdk.util.construction.CoderTranslation.toProto(CoderTranslation.java:118)
E at org.apache.beam.sdk.util.construction.SdkComponents.registerCoder(SdkComponents.java:284)
E at org.apache.beam.sdk.util.construction.PCollectionTranslation.toProto(PCollectionTranslation.java:35)
E at org.apache.beam.sdk.util.construction.SdkComponents.registerPCollection(SdkComponents.java:239)
E at org.apache.beam.sdk.util.construction.PTransformTranslation.translateAppliedPTransform(PTransformTranslation.java:610)
E at org.apache.beam.sdk.util.construction.ParDoTranslation$ParDoTranslator.translate(ParDoTranslation.java:184)
E at org.apache.beam.sdk.util.construction.PTransformTranslation.toProto(PTransformTranslation.java:277)
E at org.apache.beam.sdk.util.construction.SdkComponents.registerPTransform(SdkComponents.java:183)
E at org.apache.beam.sdk.util.construction.PipelineTranslation$1.visitPrimitiveTransform(PipelineTranslation.java:96)
E at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:593)
E at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:585)
E at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:585)
E at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:585)
E at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:585)
E at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:585)
E at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:585)
E at org.apache.beam.sdk.runners.TransformHierarchy$Node.access$500(TransformHierarchy.java:240)
E at org.apache.beam.sdk.runners.TransformHierarchy.visit(TransformHierarchy.java:214)
E at org.apache.beam.sdk.Pipeline.traverseTopologically(Pipeline.java:477)
E at org.apache.beam.sdk.util.construction.PipelineTranslation.toProto(PipelineTranslation.java:68)
E at org.apache.beam.sdk.util.construction.PipelineTranslation.toProto(PipelineTranslation.java:59)
E at org.apache.beam.sdk.util.construction.PipelineTranslation.toProto(PipelineTranslation.java:52)
E at org.apache.beam.sdk.expansion.service.ExpansionService.expand(ExpansionService.java:701)
E at org.apache.beam.sdk.expansion.service.ExpansionService.expand(ExpansionService.java:762)
E at org.apache.beam.model.expansion.v1.ExpansionServiceGrpc$MethodHandlers.invoke(ExpansionServiceGrpc.java:306)
E at org.apache.beam.vendor.grpc.v1p60p1.io.grpc.stub.ServerCalls$UnaryServerCallHandler$UnaryServerCallListener.onHalfClose(ServerCalls.java:182)
E at org.apache.beam.vendor.grpc.v1p60p1.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:351)
E at org.apache.beam.vendor.grpc.v1p60p1.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:861)
E at org.apache.beam.vendor.grpc.v1p60p1.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
E at org.apache.beam.vendor.grpc.v1p60p1.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:133)
E at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
E at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
E at java.lang.Thread.run(Thread.java:750)
E Caused by: java.io.NotSerializableException: org.apache.beam.sdk.schemas.utils.ByteBuddyUtils$TransformingMap
E at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1184)
E at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
E at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
E at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
E at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
E at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
E at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
E at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
E at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
E at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
E at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
E at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
E at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
E at java.io.ObjectOutputStream.defaultWriteFields(ObjectOutputStream.java:1548)
E at java.io.ObjectOutputStream.writeSerialData(ObjectOutputStream.java:1509)
E at java.io.ObjectOutputStream.writeOrdinaryObject(ObjectOutputStream.java:1432)
E at java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1178)
E at java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:348)
E at org.apache.beam.sdk.util.SerializableUtils.serializeToByteArray(SerializableUtils.java:55)
E ... 34 more
E
E
E java.lang.IllegalArgumentException: unable to serialize SchemaCoder<Schema: Fields:
E Field{name=tableIdentifierString, description=, type=STRING NOT NULL, options={{}}}
E Field{name=serializableDataFile, description=, type=ROW<fileFormat STRING NOT NULL, recordCount INT64 NOT NULL, fileSizeInBytes INT64 NOT NULL, splitOffsets ARRAY<INT64 NOT NULL>, keyMetadata BYTES, partitionPath STRING NOT NULL, partitionSpecId INT32 NOT NULL, columnSizes MAP<INT32 NOT NULL, INT64 NOT NULL>, valueCounts MAP<INT32 NOT NULL, INT64 NOT NULL>, nullValueCounts MAP<INT32 NOT NULL, INT64 NOT NULL>, nanValueCounts MAP<INT32 NOT NULL, INT64 NOT NULL>, upperBounds MAP<INT32 NOT NULL, BYTES NOT NULL>, lowerBounds MAP<INT32 NOT NULL, BYTES NOT NULL>, path STRING NOT NULL> NOT NULL, options={{}}}
E Encoding positions:
E {tableIdentifierString=0, serializableDataFile=1}
E Options:{{}}UUID: f534fce0-5357-451c-8993-8c5fb3b75bca UUID: f534fce0-5357-451c-8993-8c5fb3b75bca delegateCoder: org.apache.beam.sdk.coders.Coder$ByteBuddy$mdNrLYeU@2f2de4a7
apache_beam/transforms/external.py:754: RuntimeError
Check notice on line 0 in .github
github-actions / Test Results
22 skipped tests found
There are 22 skipped tests, see "Raw output" for the full list of skipped tests.
Raw output
apache_beam.examples.ml_transform.ml_transform_it_test
apache_beam.examples.snippets.transforms.elementwise.mltransform_test
apache_beam.examples.snippets.transforms.elementwise.runinference_test
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_null_key
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_populated_key
apache_beam.ml.inference.huggingface_inference_it_test
apache_beam.ml.inference.huggingface_inference_test
apache_beam.ml.inference.onnx_inference_test
apache_beam.ml.inference.pytorch_inference_test
apache_beam.ml.inference.tensorflow_inference_test
apache_beam.ml.inference.tensorrt_inference_test
apache_beam.ml.inference.vertex_ai_inference_it_test
apache_beam.ml.inference.xgboost_inference_test
apache_beam.ml.transforms.handlers_test
apache_beam.ml.transforms.tft_test
apache_beam.runners.dask.dask_runner_test
apache_beam.testing.analyzers.perf_analysis_test
apache_beam.testing.benchmarks.cloudml.cloudml_benchmark_test
apache_beam.transforms.enrichment_handlers.feast_feature_store_it_test
apache_beam.transforms.enrichment_handlers.feast_feature_store_test
apache_beam.typehints.pytorch_type_compatibility_test
apache_beam.yaml.yaml_ml_test
Check notice on line 0 in .github
github-actions / Test Results
25 tests found
There are 25 tests, see "Raw output" for the full list of tests.
Raw output
apache_beam.examples.ml_transform.ml_transform_it_test
apache_beam.examples.snippets.transforms.elementwise.mltransform_test
apache_beam.examples.snippets.transforms.elementwise.runinference_test
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_null_key
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_populated_key
apache_beam.ml.inference.huggingface_inference_it_test
apache_beam.ml.inference.huggingface_inference_test
apache_beam.ml.inference.onnx_inference_test
apache_beam.ml.inference.pytorch_inference_test
apache_beam.ml.inference.tensorflow_inference_test
apache_beam.ml.inference.tensorrt_inference_test
apache_beam.ml.inference.vertex_ai_inference_it_test
apache_beam.ml.inference.xgboost_inference_test
apache_beam.ml.transforms.handlers_test
apache_beam.ml.transforms.tft_test
apache_beam.runners.dask.dask_runner_test
apache_beam.testing.analyzers.perf_analysis_test
apache_beam.testing.benchmarks.cloudml.cloudml_benchmark_test
apache_beam.transforms.enrichment_handlers.feast_feature_store_it_test
apache_beam.transforms.enrichment_handlers.feast_feature_store_test
apache_beam.transforms.external_transform_provider_it_test.ExternalTransformProviderIT ‑ test_generate_sequence_signature_and_doc
apache_beam.transforms.external_transform_provider_it_test.ExternalTransformProviderIT ‑ test_run_generate_sequence
apache_beam.transforms.managed_iceberg_it_test.ManagedIcebergIT ‑ test_write_read_pipeline
apache_beam.typehints.pytorch_type_compatibility_test
apache_beam.yaml.yaml_ml_test
Loading