[Failing Test]: Python PostCommit is continuously failing for tf_inference_it_test #1649
7 fail, 28 skipped, 147 pass in 12h 3m 38s
182 tests 147 ✅ 12h 3m 38s ⏱️
45 suites 28 💤
45 files 7 ❌
Results for commit 289bbc4.
Annotations
github-actions / Test Results
testSimpleWrite (org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT) failed
runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformLegacyWorkerIntegrationTest/TEST-org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT.xml [took 5m 25s]
Raw output
java.lang.RuntimeException: java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
Workflow failed. Causes: S02:Create.Values/Read(CreateSource)+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/element-count+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/PrepareWrite/ParDo(Anonymous)+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/rewindowIntoGlobal/Window.Assign+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/Convert/Convert to message+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/Error on failed inserts+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/post-write+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Write Records+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/post-write+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/Error on failed inserts+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/Window.Into()/Window.Assign+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/GroupByKey/Reify+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/GroupByKey/Write failed., The job failed because a work item has failed 4 times. Look in previous log entries for the cause of each one of the 4 failures. If the logs only contain generic timeout errors related to accessing external resources, such as MongoDB, verify that the worker service account has permission to access the resource's subnetwork. For more information, see https://cloud.google.com/dataflow/docs/guides/common-errors. The work item was attempted on these workers:
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130527-gdw3-harness-kmhr,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130527-gdw3-harness-kmhr,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130527-gdw3-harness-kmhr,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130527-gdw3-harness-kmhr
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:149)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:101)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:56)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:324)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:398)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:335)
at org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT.testSimpleWrite(BigQueryStorageWriteApiSchemaTransformProviderIT.java:180)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:323)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:112)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:40)
at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:60)
at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:52)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33)
at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94)
at com.sun.proxy.$Proxy2.processTestClass(Unknown Source)
at org.gradle.api.internal.tasks.testing.worker.TestWorker$2.run(TestWorker.java:176)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.executeAndMaintainThreadName(TestWorker.java:129)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:100)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:60)
at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:56)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:113)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:65)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.run(GradleWorkerMain.java:69)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.main(GradleWorkerMain.java:74)
github-actions / Test Results
testFailedRows (org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT) failed
runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformLegacyWorkerIntegrationTest/TEST-org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT.xml [took 5m 36s]
Raw output
java.lang.RuntimeException: java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
Workflow failed. Causes: S04:Create.Values/Read(CreateSource)+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/element-count+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/PrepareWrite/ParDo(Anonymous)+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/rewindowIntoGlobal/Window.Assign+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/Convert/Convert to message+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/post-write+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/Construct failed rows and errors/Map+ExtractFailedRows/Map+PAssert$9/GroupGlobally/Reify.Window/ParDo(Anonymous)+PAssert$9/GroupGlobally/ParDo(ToSingletonIterables)+PAssert$9/GroupGlobally/WithKeys/AddKeys/Map+PAssert$9/GroupGlobally/GroupByKey/Reify+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/error-count+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Write Records+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/post-write+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/Construct failed rows and errors/Map+ExtractFailedRows/Map+PAssert$9/GroupGlobally/Reify.Window/ParDo(Anonymous)+PAssert$9/GroupGlobally/ParDo(ToSingletonIterables)+PAssert$9/GroupGlobally/WithKeys/AddKeys/Map+PAssert$9/GroupGlobally/GroupByKey/Reify+PAssert$9/GroupGlobally/GroupByKey/Session/Flatten+PAssert$9/GroupGlobally/GroupByKey/Write+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/error-count+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/Window.Into()/Window.Assign+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/GroupByKey/Reify+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/GroupByKey/Write failed., The job failed because a work item has failed 4 times. Look in previous log entries for the cause of each one of the 4 failures. If the logs only contain generic timeout errors related to accessing external resources, such as MongoDB, verify that the worker service account has permission to access the resource's subnetwork. For more information, see https://cloud.google.com/dataflow/docs/guides/common-errors. The work item was attempted on these workers:
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130533-6qyt-harness-h2t1,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130533-6qyt-harness-h2t1,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130533-6qyt-harness-h2t1,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130533-6qyt-harness-h2t1
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:149)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:101)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:56)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:324)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:398)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:335)
at org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT.testFailedRows(BigQueryStorageWriteApiSchemaTransformProviderIT.java:298)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:323)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:112)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:40)
at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:60)
at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:52)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33)
at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94)
at com.sun.proxy.$Proxy2.processTestClass(Unknown Source)
at org.gradle.api.internal.tasks.testing.worker.TestWorker$2.run(TestWorker.java:176)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.executeAndMaintainThreadName(TestWorker.java:129)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:100)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:60)
at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:56)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:113)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:65)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.run(GradleWorkerMain.java:69)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.main(GradleWorkerMain.java:74)
github-actions / Test Results
testInputElementCount (org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT) failed
runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformLegacyWorkerIntegrationTest/TEST-org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT.xml [took 5m 31s]
Raw output
java.lang.RuntimeException: java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
Workflow failed. Causes: S02:Create.Values/Read(CreateSource)+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/element-count+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/PrepareWrite/ParDo(Anonymous)+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/rewindowIntoGlobal/Window.Assign+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/Convert/Convert to message+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/Error on failed inserts+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/post-write+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Write Records+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/post-write+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/Error on failed inserts+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/Window.Into()/Window.Assign+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/GroupByKey/Reify+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/GroupByKey/Write failed., The job failed because a work item has failed 4 times. Look in previous log entries for the cause of each one of the 4 failures. If the logs only contain generic timeout errors related to accessing external resources, such as MongoDB, verify that the worker service account has permission to access the resource's subnetwork. For more information, see https://cloud.google.com/dataflow/docs/guides/common-errors. The work item was attempted on these workers:
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130538-9ep8-harness-8355,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130538-9ep8-harness-8355,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130538-9ep8-harness-8355,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130538-9ep8-harness-8355
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:149)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:101)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:56)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:324)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:398)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:335)
at org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT.testInputElementCount(BigQueryStorageWriteApiSchemaTransformProviderIT.java:231)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:323)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:112)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:40)
at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:60)
at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:52)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33)
at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94)
at com.sun.proxy.$Proxy2.processTestClass(Unknown Source)
at org.gradle.api.internal.tasks.testing.worker.TestWorker$2.run(TestWorker.java:176)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.executeAndMaintainThreadName(TestWorker.java:129)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:100)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:60)
at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:56)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:113)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:65)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.run(GradleWorkerMain.java:69)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.main(GradleWorkerMain.java:74)
github-actions / Test Results
testErrorCount (org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT) failed
runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformLegacyWorkerIntegrationTest/TEST-org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT.xml [took 5m 14s]
Raw output
java.lang.RuntimeException: java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
Workflow failed. Causes: S02:Create.Values/Read(CreateSource)+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/element-count+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/PrepareWrite/ParDo(Anonymous)+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/rewindowIntoGlobal/Window.Assign+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/Convert/Convert to message+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/post-write+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/Construct failed rows and errors/Map+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/error-count+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Write Records+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/post-write+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/Construct failed rows and errors/Map+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/error-count+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/Window.Into()/Window.Assign+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/GroupByKey/Reify+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/GroupByKey/Write failed., The job failed because a work item has failed 4 times. Look in previous log entries for the cause of each one of the 4 failures. If the logs only contain generic timeout errors related to accessing external resources, such as MongoDB, verify that the worker service account has permission to access the resource's subnetwork. For more information, see https://cloud.google.com/dataflow/docs/guides/common-errors. The work item was attempted on these workers:
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130544-mxy0-harness-24qv,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130544-mxy0-harness-24qv,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130544-mxy0-harness-24qv,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130544-mxy0-harness-24qv
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:149)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:101)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:56)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:324)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:398)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:335)
at org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT.testErrorCount(BigQueryStorageWriteApiSchemaTransformProviderIT.java:324)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:323)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:112)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:40)
at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:60)
at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:52)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33)
at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94)
at com.sun.proxy.$Proxy2.processTestClass(Unknown Source)
at org.gradle.api.internal.tasks.testing.worker.TestWorker$2.run(TestWorker.java:176)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.executeAndMaintainThreadName(TestWorker.java:129)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:100)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:60)
at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:56)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:113)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:65)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.run(GradleWorkerMain.java:69)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.main(GradleWorkerMain.java:74)
github-actions / Test Results
testWriteToDynamicDestinations (org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT) failed
runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformLegacyWorkerIntegrationTest/TEST-org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT.xml [took 5m 29s]
Raw output
java.lang.RuntimeException: java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
java.lang.NullPointerException
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.getTableContainer(FakeDatasetService.java:289)
at org.apache.beam.sdk.io.gcp.testing.FakeDatasetService.createWriteStream(FakeDatasetService.java:575)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.lambda$getOrCreateStreamName$0(StorageApiWriteUnshardedRecords.java:362)
at org.apache.beam.sdk.io.gcp.bigquery.CreateTableHelpers.createTableWrapper(CreateTableHelpers.java:70)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.getOrCreateStreamName(StorageApiWriteUnshardedRecords.java:357)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn$DestinationState.flush(StorageApiWriteUnshardedRecords.java:629)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.flushAll(StorageApiWriteUnshardedRecords.java:956)
at org.apache.beam.sdk.io.gcp.bigquery.StorageApiWriteUnshardedRecords$WriteRecordsDoFn.finishBundle(StorageApiWriteUnshardedRecords.java:1113)
Workflow failed. Causes: S02:Create.Values/Read(CreateSource)+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/element-count+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/PrepareWrite/ParDo(Anonymous)+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/rewindowIntoGlobal/Window.Assign+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/Convert/Convert to message+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/Error on failed inserts+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/post-write+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Write Records+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/post-write+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/Error on failed inserts+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/Window.Into()/Window.Assign+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/GroupByKey/Reify+BigQueryStorageWriteApiSchemaTransformProvider.BigQueryStorageWriteApiSchemaTransform/BigQueryIO.Write/StorageApiLoads/StorageApiWriteUnsharded/Reshuffle/GroupByKey/Write failed., The job failed because a work item has failed 4 times. Look in previous log entries for the cause of each one of the 4 failures. If the logs only contain generic timeout errors related to accessing external resources, such as MongoDB, verify that the worker service account has permission to access the resource's subnetwork. For more information, see https://cloud.google.com/dataflow/docs/guides/common-errors. The work item was attempted on these workers:
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130549-840b-harness-9gkq,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130549-840b-harness-9gkq,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130549-840b-harness-9gkq,
Root cause: Work item failed.
Worker ID: bigquerystoragewriteapisc-03130549-840b-harness-9gkq
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:149)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:101)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:56)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:324)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:398)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:335)
at org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT.testWriteToDynamicDestinations(BigQueryStorageWriteApiSchemaTransformProviderIT.java:208)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:323)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:112)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:40)
at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:60)
at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:52)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33)
at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94)
at com.sun.proxy.$Proxy2.processTestClass(Unknown Source)
at org.gradle.api.internal.tasks.testing.worker.TestWorker$2.run(TestWorker.java:176)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.executeAndMaintainThreadName(TestWorker.java:129)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:100)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:60)
at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:56)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:113)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:65)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.run(GradleWorkerMain.java:69)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.main(GradleWorkerMain.java:74)
github-actions / Test Results
testDirectRead (org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProviderIT) failed
runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformLegacyWorkerIntegrationTest/TEST-org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProviderIT.xml [took 9s]
Raw output
java.lang.RuntimeException: Failed to create a workflow job: Consumer 'projects/parent-project' is invalid: Resource projects/parent-project could not be found..
at org.apache.beam.runners.dataflow.DataflowRunner.run(DataflowRunner.java:1467)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:109)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:101)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:56)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:324)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:398)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:335)
at org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProviderIT.testDirectRead(BigQueryDirectReadSchemaTransformProviderIT.java:284)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:323)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:112)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:40)
at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:60)
at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:52)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33)
at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94)
at com.sun.proxy.$Proxy2.processTestClass(Unknown Source)
at org.gradle.api.internal.tasks.testing.worker.TestWorker$2.run(TestWorker.java:176)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.executeAndMaintainThreadName(TestWorker.java:129)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:100)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:60)
at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:56)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:113)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:65)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.run(GradleWorkerMain.java:69)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.main(GradleWorkerMain.java:74)
Caused by: com.google.api.client.googleapis.json.GoogleJsonResponseException: 400 Bad Request
POST https://dataflow.googleapis.com/v1b3/projects/parent-project/locations/us-central1/jobs
{
"code" : 400,
"details" : [ {
"@type" : "type.googleapis.com/google.rpc.Help"
}, {
"@type" : "type.googleapis.com/google.rpc.ErrorInfo",
"reason" : "CONSUMER_INVALID"
} ],
"errors" : [ {
"domain" : "global",
"message" : "Consumer 'projects/parent-project' is invalid: Resource projects/parent-project could not be found..",
"reason" : "badRequest"
} ],
"message" : "Consumer 'projects/parent-project' is invalid: Resource projects/parent-project could not be found..",
"status" : "INVALID_ARGUMENT"
}
at com.google.api.client.googleapis.json.GoogleJsonResponseException.from(GoogleJsonResponseException.java:146)
at com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:118)
at com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:37)
at com.google.api.client.googleapis.services.AbstractGoogleClientRequest$1.interceptResponse(AbstractGoogleClientRequest.java:439)
at com.google.api.client.http.HttpRequest.execute(HttpRequest.java:1111)
at com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:525)
at com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:466)
at com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:576)
at org.apache.beam.runners.dataflow.DataflowClient.createJob(DataflowClient.java:64)
at org.apache.beam.runners.dataflow.DataflowRunner.run(DataflowRunner.java:1453)
... 52 more
github-actions / Test Results
testDirectReadWithSelectedFieldsAndRowRestriction (org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProviderIT) failed
runners/google-cloud-dataflow-java/build/test-results/googleCloudPlatformLegacyWorkerIntegrationTest/TEST-org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProviderIT.xml [took 8s]
Raw output
java.lang.RuntimeException: Failed to create a workflow job: Consumer 'projects/parent-project' is invalid: Resource projects/parent-project could not be found..
at org.apache.beam.runners.dataflow.DataflowRunner.run(DataflowRunner.java:1467)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:109)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:101)
at org.apache.beam.runners.dataflow.TestDataflowRunner.run(TestDataflowRunner.java:56)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:324)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:398)
at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:335)
at org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProviderIT.testDirectReadWithSelectedFieldsAndRowRestriction(BigQueryDirectReadSchemaTransformProviderIT.java:348)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:323)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:112)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58)
at org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:40)
at org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:60)
at org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:52)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
at org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
at org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33)
at org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94)
at com.sun.proxy.$Proxy2.processTestClass(Unknown Source)
at org.gradle.api.internal.tasks.testing.worker.TestWorker$2.run(TestWorker.java:176)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.executeAndMaintainThreadName(TestWorker.java:129)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:100)
at org.gradle.api.internal.tasks.testing.worker.TestWorker.execute(TestWorker.java:60)
at org.gradle.process.internal.worker.child.ActionExecutionWorker.execute(ActionExecutionWorker.java:56)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:113)
at org.gradle.process.internal.worker.child.SystemApplicationClassLoaderWorker.call(SystemApplicationClassLoaderWorker.java:65)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.run(GradleWorkerMain.java:69)
at worker.org.gradle.process.internal.worker.GradleWorkerMain.main(GradleWorkerMain.java:74)
Caused by: com.google.api.client.googleapis.json.GoogleJsonResponseException: 400 Bad Request
POST https://dataflow.googleapis.com/v1b3/projects/parent-project/locations/us-central1/jobs
{
"code" : 400,
"details" : [ {
"@type" : "type.googleapis.com/google.rpc.Help"
}, {
"@type" : "type.googleapis.com/google.rpc.ErrorInfo",
"reason" : "CONSUMER_INVALID"
} ],
"errors" : [ {
"domain" : "global",
"message" : "Consumer 'projects/parent-project' is invalid: Resource projects/parent-project could not be found..",
"reason" : "badRequest"
} ],
"message" : "Consumer 'projects/parent-project' is invalid: Resource projects/parent-project could not be found..",
"status" : "INVALID_ARGUMENT"
}
at com.google.api.client.googleapis.json.GoogleJsonResponseException.from(GoogleJsonResponseException.java:146)
at com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:118)
at com.google.api.client.googleapis.services.json.AbstractGoogleJsonClientRequest.newExceptionOnError(AbstractGoogleJsonClientRequest.java:37)
at com.google.api.client.googleapis.services.AbstractGoogleClientRequest$1.interceptResponse(AbstractGoogleClientRequest.java:439)
at com.google.api.client.http.HttpRequest.execute(HttpRequest.java:1111)
at com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:525)
at com.google.api.client.googleapis.services.AbstractGoogleClientRequest.executeUnparsed(AbstractGoogleClientRequest.java:466)
at com.google.api.client.googleapis.services.AbstractGoogleClientRequest.execute(AbstractGoogleClientRequest.java:576)
at org.apache.beam.runners.dataflow.DataflowClient.createJob(DataflowClient.java:64)
at org.apache.beam.runners.dataflow.DataflowRunner.run(DataflowRunner.java:1453)
... 52 more
Check notice on line 0 in .github
github-actions / Test Results
28 skipped tests found
There are 28 skipped tests, see "Raw output" for the full list of skipped tests.
Raw output
org.apache.beam.sdk.io.gcp.bigquery.FileLoadsStreamingIT ‑ testDynamicDestinationsWithAutoShardingAndCopyJobs[0]
org.apache.beam.sdk.io.gcp.bigquery.FileLoadsStreamingIT ‑ testDynamicDestinationsWithFixedShards[0]
org.apache.beam.sdk.io.gcp.bigquery.FileLoadsStreamingIT ‑ testLoadWithAutoShardingAndCopyJobs[0]
org.apache.beam.sdk.io.gcp.bigquery.FileLoadsStreamingIT ‑ testLoadWithFixedShards[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkDefaultValuesIT ‑ testMissingValueSchemaUnknownTakeNull
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithAutoSchemaUpdate[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithAutoSchemaUpdate[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithAutoSchemaUpdate[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithIgnoreUnknownValues[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithIgnoreUnknownValues[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithIgnoreUnknownValues[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithIgnoreUnknownValues[3]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnce[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnce[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnce[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnce[3]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithAutoSchemaUpdate[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithAutoSchemaUpdate[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithAutoSchemaUpdate[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithIgnoreUnknownValues[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithIgnoreUnknownValues[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithIgnoreUnknownValues[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithIgnoreUnknownValues[3]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnce[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnce[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnce[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnce[3]
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT ‑ testReadWithDataBoost
Check notice on line 0 in .github
github-actions / Test Results
182 tests found
There are 182 tests, see "Raw output" for the full list of tests.
Raw output
org.apache.beam.sdk.io.gcp.bigquery.BigQueryClusteringIT ‑ testE2EBigQueryClusteringNoPartitionDynamicDestinations
org.apache.beam.sdk.io.gcp.bigquery.BigQueryClusteringIT ‑ testE2EBigQueryClusteringNoPartitionTableFunction
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOJsonIT ‑ testFileLoadWriteExportRead
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOJsonIT ‑ testLegacyStreamingWriteDefaultRead
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOJsonIT ‑ testQueryRead
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOJsonIT ‑ testStorageWriteRead
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageQueryIT ‑ testBigQueryStorageQuery1G
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageQueryIT ‑ testBigQueryStorageQueryWithErrorHandling1M
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageReadIT ‑ testBigQueryStorageRead1GArrow
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageReadIT ‑ testBigQueryStorageRead1GAvro
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageReadIT ‑ testBigQueryStorageRead1MErrorHandlingArrow
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageReadIT ‑ testBigQueryStorageRead1MErrorHandlingAvro
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageReadIT ‑ testBigQueryStorageReadProjectionPushdown
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageReadIT ‑ testBigQueryStorageReadWithArrow
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageReadIT ‑ testBigQueryStorageReadWithAvro
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageWriteIT ‑ testBigQueryStorageWrite3KProtoALOStreaming
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageWriteIT ‑ testBigQueryStorageWrite3KProtoStreaming
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageWriteIT ‑ testBigQueryStorageWrite3MProto
org.apache.beam.sdk.io.gcp.bigquery.BigQueryIOStorageWriteIT ‑ testBigQueryStorageWrite3MProtoALO
org.apache.beam.sdk.io.gcp.bigquery.BigQueryNestedRecordsIT ‑ testNestedRecords
org.apache.beam.sdk.io.gcp.bigquery.BigQuerySchemaUpdateOptionsIT ‑ runWriteTestTempTableAndDynamicDestination
org.apache.beam.sdk.io.gcp.bigquery.BigQuerySchemaUpdateOptionsIT ‑ testAllowFieldAddition
org.apache.beam.sdk.io.gcp.bigquery.BigQuerySchemaUpdateOptionsIT ‑ testAllowFieldRelaxation
org.apache.beam.sdk.io.gcp.bigquery.BigQueryTimePartitioningClusteringIT ‑ testE2EBigQueryClustering
org.apache.beam.sdk.io.gcp.bigquery.BigQueryTimePartitioningClusteringIT ‑ testE2EBigQueryClusteringDynamicDestinations
org.apache.beam.sdk.io.gcp.bigquery.BigQueryTimePartitioningClusteringIT ‑ testE2EBigQueryClusteringTableFunction
org.apache.beam.sdk.io.gcp.bigquery.BigQueryTimePartitioningClusteringIT ‑ testE2EBigQueryTimePartitioning
org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT ‑ testLegacyQueryWithoutReshuffle
org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT ‑ testNewTypesQueryWithReshuffle
org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT ‑ testNewTypesQueryWithoutReshuffle
org.apache.beam.sdk.io.gcp.bigquery.BigQueryToTableIT ‑ testStandardQueryWithoutCustom
org.apache.beam.sdk.io.gcp.bigquery.FileLoadsStreamingIT ‑ testDynamicDestinationsWithAutoShardingAndCopyJobs[0]
org.apache.beam.sdk.io.gcp.bigquery.FileLoadsStreamingIT ‑ testDynamicDestinationsWithAutoShardingAndCopyJobs[1]
org.apache.beam.sdk.io.gcp.bigquery.FileLoadsStreamingIT ‑ testDynamicDestinationsWithFixedShards[0]
org.apache.beam.sdk.io.gcp.bigquery.FileLoadsStreamingIT ‑ testDynamicDestinationsWithFixedShards[1]
org.apache.beam.sdk.io.gcp.bigquery.FileLoadsStreamingIT ‑ testLoadWithAutoShardingAndCopyJobs[0]
org.apache.beam.sdk.io.gcp.bigquery.FileLoadsStreamingIT ‑ testLoadWithAutoShardingAndCopyJobs[1]
org.apache.beam.sdk.io.gcp.bigquery.FileLoadsStreamingIT ‑ testLoadWithFixedShards[0]
org.apache.beam.sdk.io.gcp.bigquery.FileLoadsStreamingIT ‑ testLoadWithFixedShards[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiDirectWriteProtosIT ‑ testDirectWriteProtos[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiDirectWriteProtosIT ‑ testDirectWriteProtos[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiDirectWriteProtosIT ‑ testDirectWriteProtos[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiDirectWriteProtosIT ‑ testDirectWriteProtos[3]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkCreateIfNeededIT ‑ testCreateManyTables[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkCreateIfNeededIT ‑ testCreateManyTables[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkDefaultValuesIT ‑ testMissingRequiredValueSchemaKnownTakeDefault
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkDefaultValuesIT ‑ testMissingRequiredValueSchemaKnownTakeNull
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkDefaultValuesIT ‑ testMissingRequiredValueSchemaUnknownTakeDefault
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkDefaultValuesIT ‑ testMissingValueSchemaKnownTakeDefault
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkDefaultValuesIT ‑ testMissingValueSchemaKnownTakeNull
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkDefaultValuesIT ‑ testMissingValueSchemaUnknownTakeDefault
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkDefaultValuesIT ‑ testMissingValueSchemaUnknownTakeNull
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkFailedRowsIT ‑ testInvalidRowCaughtByBigquery[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkFailedRowsIT ‑ testInvalidRowCaughtByBigquery[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkFailedRowsIT ‑ testInvalidRowCaughtByBigquery[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkFailedRowsIT ‑ testInvalidRowCaughtByBigquery[3]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkFailedRowsIT ‑ testSchemaMismatchCaughtByBeam[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkFailedRowsIT ‑ testSchemaMismatchCaughtByBeam[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkFailedRowsIT ‑ testSchemaMismatchCaughtByBeam[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkFailedRowsIT ‑ testSchemaMismatchCaughtByBeam[3]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkRowUpdateIT ‑ testCdc
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithAutoSchemaUpdate[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithAutoSchemaUpdate[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithAutoSchemaUpdate[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithAutoSchemaUpdate[3]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithIgnoreUnknownValues[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithIgnoreUnknownValues[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithIgnoreUnknownValues[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnceWithIgnoreUnknownValues[3]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnce[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnce[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnce[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testAtLeastOnce[3]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithAutoSchemaUpdate[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithAutoSchemaUpdate[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithAutoSchemaUpdate[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithAutoSchemaUpdate[3]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithIgnoreUnknownValues[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithIgnoreUnknownValues[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithIgnoreUnknownValues[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnceWithIgnoreUnknownValues[3]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnce[0]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnce[1]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnce[2]
org.apache.beam.sdk.io.gcp.bigquery.StorageApiSinkSchemaUpdateIT ‑ testExactlyOnce[3]
org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProtoIT ‑ testBaseTableRow
org.apache.beam.sdk.io.gcp.bigquery.TableRowToStorageApiProtoIT ‑ testNestedRichTypesAndNull
org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProviderIT ‑ testDirectRead
org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProviderIT ‑ testDirectReadWithSelectedFieldsAndRowRestriction
org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryDirectReadSchemaTransformProviderIT ‑ testValidateConfig
org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT ‑ testErrorCount
org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT ‑ testFailedRows
org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT ‑ testInputElementCount
org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT ‑ testInvalidConfig
org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT ‑ testSimpleWrite
org.apache.beam.sdk.io.gcp.bigquery.providers.BigQueryStorageWriteApiSchemaTransformProviderIT ‑ testWriteToDynamicDestinations
org.apache.beam.sdk.io.gcp.bigtable.BigtableReadIT ‑ testE2EBigtableRead
org.apache.beam.sdk.io.gcp.bigtable.BigtableReadIT ‑ testE2EBigtableSegmentRead
org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProviderIT ‑ testInvalidConfigs
org.apache.beam.sdk.io.gcp.bigtable.BigtableReadSchemaTransformProviderIT ‑ testRead
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteIT ‑ testE2EBigtableWrite
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteIT ‑ testE2EBigtableWriteWithEmptyMutationFailures
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteIT ‑ testE2EBigtableWriteWithEmptyRowFailures
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteIT ‑ testE2EBigtableWriteWithInvalidColumnFamilyFailures
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteIT ‑ testE2EBigtableWriteWithInvalidTimestampFailures
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteIT ‑ testE2EBigtableWriteWithOversizedQualifierFailures
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProviderIT ‑ testDeleteCellsFromColumn
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProviderIT ‑ testDeleteCellsFromColumnWithTimestampRange
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProviderIT ‑ testDeleteColumnFamily
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProviderIT ‑ testDeleteRow
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProviderIT ‑ testInvalidConfigs
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProviderIT ‑ testSetMutationNewColumn
org.apache.beam.sdk.io.gcp.bigtable.BigtableWriteSchemaTransformProviderIT ‑ testSetMutationsExistingColumn
org.apache.beam.sdk.io.gcp.bigtable.changestreams.it.BigtableChangeStreamIT ‑ testComplexMutation
org.apache.beam.sdk.io.gcp.bigtable.changestreams.it.BigtableChangeStreamIT ‑ testDeleteCell
org.apache.beam.sdk.io.gcp.bigtable.changestreams.it.BigtableChangeStreamIT ‑ testDeleteColumnFamily
org.apache.beam.sdk.io.gcp.bigtable.changestreams.it.BigtableChangeStreamIT ‑ testDeleteRow
org.apache.beam.sdk.io.gcp.bigtable.changestreams.it.BigtableChangeStreamIT ‑ testLargeMutation
org.apache.beam.sdk.io.gcp.bigtable.changestreams.it.BigtableChangeStreamIT ‑ testManyMutations
org.apache.beam.sdk.io.gcp.bigtable.changestreams.it.BigtableChangeStreamIT ‑ testReadBigtableChangeStream
org.apache.beam.sdk.io.gcp.datastore.SplitQueryFnIT ‑ testSplitQueryFnWithLargeDataset
org.apache.beam.sdk.io.gcp.datastore.SplitQueryFnIT ‑ testSplitQueryFnWithSmallDataset
org.apache.beam.sdk.io.gcp.datastore.V1ReadIT ‑ testE2EV1Read
org.apache.beam.sdk.io.gcp.datastore.V1ReadIT ‑ testE2EV1ReadWithGQLQueryWithLimit
org.apache.beam.sdk.io.gcp.datastore.V1ReadIT ‑ testE2EV1ReadWithGQLQueryWithNoLimit
org.apache.beam.sdk.io.gcp.datastore.V1WriteIT ‑ testDatastoreWriterFnWithDuplicatedEntities
org.apache.beam.sdk.io.gcp.datastore.V1WriteIT ‑ testE2EV1Write
org.apache.beam.sdk.io.gcp.datastore.V1WriteIT ‑ testE2EV1WriteWithLargeEntities
org.apache.beam.sdk.io.gcp.firestore.it.FirestoreV1IT ‑ batchGet
org.apache.beam.sdk.io.gcp.firestore.it.FirestoreV1IT ‑ batchWrite_partialFailureOutputsToDeadLetterQueue
org.apache.beam.sdk.io.gcp.firestore.it.FirestoreV1IT ‑ listCollections
org.apache.beam.sdk.io.gcp.firestore.it.FirestoreV1IT ‑ listDocuments
org.apache.beam.sdk.io.gcp.firestore.it.FirestoreV1IT ‑ partitionQuery
org.apache.beam.sdk.io.gcp.firestore.it.FirestoreV1IT ‑ runQuery
org.apache.beam.sdk.io.gcp.firestore.it.FirestoreV1IT ‑ write
org.apache.beam.sdk.io.gcp.healthcare.FhirIOLROIT ‑ test_FhirIO_deidentify
org.apache.beam.sdk.io.gcp.healthcare.FhirIOLROIT ‑ test_FhirIO_exportFhirResources_BigQuery
org.apache.beam.sdk.io.gcp.healthcare.FhirIOLROIT ‑ test_FhirIO_exportFhirResources_Gcs
org.apache.beam.sdk.io.gcp.healthcare.FhirIOPatientEverythingIT ‑ testFhirIOPatientEverything[R4]
org.apache.beam.sdk.io.gcp.healthcare.FhirIOSearchIT ‑ testFhirIOSearchWithGenericParameters[R4]
org.apache.beam.sdk.io.gcp.healthcare.FhirIOSearchIT ‑ testFhirIOSearch[R4]
org.apache.beam.sdk.io.gcp.healthcare.FhirIOSearchIT ‑ testFhirIOSearch_emptyResult[R4]
org.apache.beam.sdk.io.gcp.healthcare.FhirIOWriteIT ‑ testFhirIO_ExecuteBundle[DSTU2]
org.apache.beam.sdk.io.gcp.healthcare.FhirIOWriteIT ‑ testFhirIO_ExecuteBundle[R4]
org.apache.beam.sdk.io.gcp.healthcare.FhirIOWriteIT ‑ testFhirIO_ExecuteBundle[STU3]
org.apache.beam.sdk.io.gcp.healthcare.FhirIOWriteIT ‑ testFhirIO_ExecuteBundle_parseResponse[DSTU2]
org.apache.beam.sdk.io.gcp.healthcare.FhirIOWriteIT ‑ testFhirIO_ExecuteBundle_parseResponse[R4]
org.apache.beam.sdk.io.gcp.healthcare.FhirIOWriteIT ‑ testFhirIO_ExecuteBundle_parseResponse[STU3]
org.apache.beam.sdk.io.gcp.healthcare.FhirIOWriteIT ‑ testFhirIO_Import[DSTU2]
org.apache.beam.sdk.io.gcp.healthcare.FhirIOWriteIT ‑ testFhirIO_Import[R4]
org.apache.beam.sdk.io.gcp.healthcare.FhirIOWriteIT ‑ testFhirIO_Import[STU3]
org.apache.beam.sdk.io.gcp.healthcare.HL7v2IOReadIT ‑ testHL7v2IO_ListHL7v2Messages
org.apache.beam.sdk.io.gcp.healthcare.HL7v2IOReadIT ‑ testHL7v2IO_ListHL7v2Messages_filtered
org.apache.beam.sdk.io.gcp.healthcare.HL7v2IOReadWriteIT ‑ testHL7v2IOE2E
org.apache.beam.sdk.io.gcp.healthcare.HL7v2IOReadWriteIT ‑ testHL7v2IOGetAllByReadParameterE2E
org.apache.beam.sdk.io.gcp.healthcare.HL7v2IOReadWriteIT ‑ testHL7v2IOGetAllE2E
org.apache.beam.sdk.io.gcp.healthcare.HL7v2IOWriteIT ‑ testHL7v2IOWrite
org.apache.beam.sdk.io.gcp.pubsub.PubsubSchemaIT ‑ testGetSchema
org.apache.beam.sdk.io.gcp.pubsub.PubsubSchemaIT ‑ testGetSchemaPath
org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteIT ‑ testBoundedWriteLargeMessage
org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteIT ‑ testBoundedWriteMessageWithAttributes
org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteIT ‑ testBoundedWriteSequence
org.apache.beam.sdk.io.gcp.pubsub.PubsubWriteIT ‑ testBoundedWriteSmallMessage
org.apache.beam.sdk.io.gcp.pubsublite.ReadWriteIT ‑ testPubsubLiteWriteReadWithSchemaTransform
org.apache.beam.sdk.io.gcp.pubsublite.ReadWriteIT ‑ testReadWrite
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT ‑ testQuery
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT ‑ testQueryWithTimeoutError
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT ‑ testQueryWithTimeoutErrorPG
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT ‑ testRead
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT ‑ testReadAllRecordsInDb
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT ‑ testReadFailsBadSession
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT ‑ testReadFailsBadTable
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT ‑ testReadWithDataBoost
org.apache.beam.sdk.io.gcp.spanner.SpannerReadIT ‑ testReadWithTimeoutError
org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT ‑ testFailFast
org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT ‑ testPgFailFast
org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT ‑ testReportFailures
org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT ‑ testSequentialWrite
org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT ‑ testWrite
org.apache.beam.sdk.io.gcp.spanner.SpannerWriteIT ‑ testWriteViaSchemaTransform
org.apache.beam.sdk.io.gcp.storage.GcsKmsKeyIT ‑ testGcsWriteWithKmsKey
org.apache.beam.sdk.io.gcp.storage.GcsMatchIT ‑ testGcsMatchContinuously