[DO NOT MERGE] Run all PostCommit and PreCommit Tests against Release Branch #4106
GitHub Actions / Test Results
failed
Dec 19, 2024 in 0s
1 fail, 19 skipped, 2 pass in 16m 2s
22 tests 2 ✅ 16m 2s ⏱️
1 suites 19 💤
1 files 1 ❌
Results for commit c1ba1bd.
Annotations
Check warning on line 0 in apache_beam.examples.wordcount_it_test.WordCountIT
github-actions / Test Results
test_wordcount_it (apache_beam.examples.wordcount_it_test.WordCountIT) failed
sdks/python/pytest-beam_python3.12_sdk.xml [took 2s]
Raw output
apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Thu, 19 Dec 2024 14:44:37 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '492', '-content-encoding': 'gzip'}>, content <{
"error": {
"code": 400,
"message": "(5e85d8be376aa2b0): The workflow could not be created. Causes: (534a96c340fd38e9): Dataflow quota error for jobs-per-project quota. Project apache-beam-testing is running 300 jobs. Please check the quota usage via GCP Console. If it exceeds the limit, please wait for a workflow to finish or contact Google Cloud Support to request an increase in quota. If it does not, contact Google Cloud Support.",
"status": "FAILED_PRECONDITION"
}
}
>
self = <apache_beam.examples.wordcount_it_test.WordCountIT testMethod=test_wordcount_it>
@pytest.mark.it_postcommit
@pytest.mark.it_validatescontainer
def test_wordcount_it(self):
> self._run_wordcount_it(wordcount.run)
apache_beam/examples/wordcount_it_test.py:50:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/examples/wordcount_it_test.py:150: in _run_wordcount_it
run_wordcount(
apache_beam/examples/wordcount.py:109: in run
result = pipeline.run()
apache_beam/pipeline.py:570: in run
self._options).run(False)
apache_beam/pipeline.py:594: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/dataflow/test_dataflow_runner.py:53: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/dataflow/dataflow_runner.py:502: in run_pipeline
self.dataflow_client.create_job(self.job), self)
apache_beam/utils/retry.py:298: in wrapper
return fun(*args, **kwargs)
apache_beam/runners/dataflow/internal/apiclient.py:725: in create_job
return self.submit_job_description(job)
apache_beam/utils/retry.py:298: in wrapper
return fun(*args, **kwargs)
apache_beam/runners/dataflow/internal/apiclient.py:831: in submit_job_description
response = self._client.projects_locations_jobs.Create(request)
apache_beam/runners/dataflow/internal/clients/dataflow/dataflow_v1b3_client.py:718: in Create
return self._RunMethod(config, request, global_params=global_params)
../../build/gradleenv/2050596100/lib/python3.12/site-packages/apitools/base/py/base_api.py:731: in _RunMethod
return self.ProcessHttpResponse(method_config, http_response, request)
../../build/gradleenv/2050596100/lib/python3.12/site-packages/apitools/base/py/base_api.py:737: in ProcessHttpResponse
self.__ProcessHttpResponse(method_config, http_response, request))
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <apache_beam.runners.dataflow.internal.clients.dataflow.dataflow_v1b3_client.DataflowV1b3.ProjectsLocationsJobsService object at 0x7fb9930a15e0>
method_config = <ApiMethodInfo
relative_path: 'v1b3/projects/{projectId}/locations/{location}/jobs'
method_id: 'dataflow.projects.lo...DataflowProjectsLocationsJobsCreateRequest'
response_type_name: 'Job'
request_field: 'job'
supports_download: False>
http_response = Response(info={'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Thu, 1...', request_url='https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json')
request = <DataflowProjectsLocationsJobsCreateRequest
job: <Job
clientRequestId: '20241219144436289060-5807'
environment: <En...empFiles: []
type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
location: 'us-central1'
projectId: 'apache-beam-testing'>
def __ProcessHttpResponse(self, method_config, http_response, request):
"""Process the given http response."""
if http_response.status_code not in (http_client.OK,
http_client.CREATED,
http_client.NO_CONTENT):
> raise exceptions.HttpError.FromResponse(
http_response, method_config=method_config, request=request)
E apitools.base.py.exceptions.HttpBadRequestError: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'vary': 'Origin, X-Origin, Referer', 'content-type': 'application/json; charset=UTF-8', 'date': 'Thu, 19 Dec 2024 14:44:37 GMT', 'server': 'ESF', 'cache-control': 'private', 'x-xss-protection': '0', 'x-frame-options': 'SAMEORIGIN', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'status': '400', 'content-length': '492', '-content-encoding': 'gzip'}>, content <{
E "error": {
E "code": 400,
E "message": "(5e85d8be376aa2b0): The workflow could not be created. Causes: (534a96c340fd38e9): Dataflow quota error for jobs-per-project quota. Project apache-beam-testing is running 300 jobs. Please check the quota usage via GCP Console. If it exceeds the limit, please wait for a workflow to finish or contact Google Cloud Support to request an increase in quota. If it does not, contact Google Cloud Support.",
E "status": "FAILED_PRECONDITION"
E }
E }
E >
../../build/gradleenv/2050596100/lib/python3.12/site-packages/apitools/base/py/base_api.py:603: HttpBadRequestError
Loading