You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The BOM builds are taking way more time (and money) than they should given the parallelization and infrastructure capacity.
A lot of steps are way slower than expected when there are hundreds of parallel jobs running on ci.jenkins.io.
This issue is about identifying the most probable causes by ruling out the Kubernetes plugin and/or the AWS EKS cloud used for agents.
rule out (or point at) the Kubernetes plugin by trying a BOM build using Azure VMs
Fine enough if that is straightforward to do, but the most direct diagnosis would be to capture a controller thread dump at the moment that a particular (sh?) step is running without the agent actually running the corresponding external process, or that a step has supposedly completed yet the next step in the pipeline has not yet started. Being able to reproduce the problem and look at a thread dump could potentially mean immediately zeroing in one some 🤦 mistake in some plugin that is easily solved once we see it.
The past months showed that moving the agent kubernetes cluster close to the controller (from AWS to Azure) did solve the problem. As such I'm closing the issue as not reproducible.
Service(s)
ci.jenkins.io
Summary
Follow up of jenkinsci/bom#1969 and #3521.
The BOM builds are taking way more time (and money) than they should given the parallelization and infrastructure capacity.
A lot of steps are way slower than expected when there are hundreds of parallel jobs running on ci.jenkins.io.
This issue is about identifying the most probable causes by ruling out the Kubernetes plugin and/or the AWS EKS cloud used for agents.
The following tasks are expected:
Reproduction steps
No response
The text was updated successfully, but these errors were encountered: