Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Report failures of periodic jobs to the cluster-api Slack channel #10520

Closed
sbueringer opened this issue Apr 26, 2024 · 23 comments
Closed

Report failures of periodic jobs to the cluster-api Slack channel #10520

sbueringer opened this issue Apr 26, 2024 · 23 comments
Assignees
Labels
area/e2e-testing Issues or PRs related to e2e testing kind/feature Categorizes issue or PR as related to a new feature. lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. needs-triage Indicates an issue or PR lacks a `triage/foo` label and requires one. priority/backlog Higher priority than priority/awaiting-more-evidence.

Comments

@sbueringer
Copy link
Member

I noticed that CAPO is reporting periodic test failures to Slack, e.g.: https://kubernetes.slack.com/archives/CFKJB65G9/p1713540048571589

I think think this is a great way to surface issues with CI (and also folks can directly start a thread based on a Slack comment like this)

This could be configured ~ like this: https://github.com/kubernetes/test-infra/blob/5d7e1db75dce28537ba5f17476882869d1b94b0a/config/jobs/kubernetes-sigs/cluster-api-provider-openstack/cluster-api-provider-openstack-periodics.yaml#L48-L55

What do you think?

@sbueringer
Copy link
Member Author

cc @chrischdi @fabriziopandini

@k8s-ci-robot
Copy link
Contributor

This issue is currently awaiting triage.

CAPI contributors will take a look as soon as possible, apply one of the triage/* labels and provide further guidance.

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository.

@k8s-ci-robot k8s-ci-robot added the needs-triage Indicates an issue or PR lacks a `triage/foo` label and requires one. label Apr 26, 2024
@sbueringer sbueringer added the area/e2e-testing Issues or PRs related to e2e testing label Apr 26, 2024
@chrischdi
Copy link
Member

Oh wow, yeah that would be a great thing. I just fear that it may pollute the channel too much. But we could try and fail fast by asking for feedback if it is too much later on in the community meeting or via a slack thread/poll.

@killianmuldoon
Copy link
Contributor

do we know if this respects testgrid-num-failures-to-alert? If so it could be great.

@sbueringer
Copy link
Member Author

I'm not sure if it respects that. We could try and rollback if it doesn't?

@sbueringer
Copy link
Member Author

If it still pollutes the channel too much after considering testgrid-num-failures-to-alert we have to focus more on CI :D

(I"m currently guessing that we would get one Slack message for every mail that we get today, but I don't know)

@killianmuldoon
Copy link
Contributor

One slack message per mail would be perfect - more would disrupt the channel

WDYT about enabling it for CAPV first?

@killianmuldoon
Copy link
Contributor

Also fine with making the change and rolling back if it doesn't work

@sbueringer
Copy link
Member Author

One slack message per mail would be perfect - more would disrupt the channel
WDYT about enabling it for CAPV first?

Fine for me, we can also ask the OpenStack folks how spamy it is for them today (cc @mdbooth @lentzi90)

@lentzi90
Copy link
Contributor

For CAPO we get a message for every failure and email only after 2 failures in a row. I think it has been tolerable for us, but that indicates it does not check testgrid-num-failures-to-alert (at least the way we have it configured)

@sbueringer
Copy link
Member Author

Hm okay, every failure is just too much. So we should probably take a closer look at the configuration / implementation. One message for every failure just doesn't make sense for the amount of tests/failures we have (the signal/noise ratio is just wrong)

@fabriziopandini
Copy link
Member

+1 to test this if we find a config reasonably noisy (but not too much noisy)
cc @kubernetes-sigs/cluster-api-release-team

/priority backlog
/kind feature

@k8s-ci-robot k8s-ci-robot added priority/backlog Higher priority than priority/awaiting-more-evidence. kind/feature Categorizes issue or PR as related to a new feature. labels Apr 29, 2024
@adilGhaffarDev
Copy link
Contributor

+1 from my side too. Tagging CI lead @Sunnatillo
I will add this to improvement tasks for v1.8 cycle. CI team can look into this one.

@Sunnatillo
Copy link
Contributor

Sounds great. I will take a look

@Sunnatillo
Copy link
Contributor

I guess this testgrid-num-failures-to-alert should help with the amount of the noise. If we set it, for example, to 5 we will be sure that we will receive messages about constantly failing tests. This makes the config to sent the alert after 5 continuous failures.

@Sunnatillo
Copy link
Contributor

/assign @Sunnatillo

@lentzi90
Copy link
Contributor

@Sunnatillo testgrid-num-failures-to-alert does not affect the slack messages for CAPO at least. Only emails are affected by that in my experience.

@Sunnatillo
Copy link
Contributor

@Sunnatillo testgrid-num-failures-to-alert does not affect the slack messages for CAPO at least. Only emails are affected by that in my experience.

Thank you for the update. I will open the issue in test-infra, try to find the way to do it.

@Sunnatillo
Copy link
Contributor

I opened an issue regarding this in test-infra:
kubernetes/test-infra#32687

@k8s-triage-robot
Copy link

The Kubernetes project currently lacks enough contributors to adequately respond to all issues.

This bot triages un-triaged issues according to the following rules:

  • After 90d of inactivity, lifecycle/stale is applied
  • After 30d of inactivity since lifecycle/stale was applied, lifecycle/rotten is applied
  • After 30d of inactivity since lifecycle/rotten was applied, the issue is closed

You can:

  • Mark this issue as fresh with /remove-lifecycle stale
  • Close this issue with /close
  • Offer to help out with Issue Triage

Please send feedback to sig-contributor-experience at kubernetes/community.

/lifecycle stale

@k8s-ci-robot k8s-ci-robot added the lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. label Sep 1, 2024
@sbueringer
Copy link
Member Author

sbueringer commented Sep 2, 2024

Maybe let's close this here until kubernetes-sigs/prow#195 has been implemented? (which might take a very long time if nobody volunteers for it)

@fabriziopandini
Copy link
Member

As per comment above
/close

@k8s-ci-robot
Copy link
Contributor

@fabriziopandini: Closing this issue.

In response to this:

As per comment above
/close

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area/e2e-testing Issues or PRs related to e2e testing kind/feature Categorizes issue or PR as related to a new feature. lifecycle/stale Denotes an issue or PR has remained open with no activity and has become stale. needs-triage Indicates an issue or PR lacks a `triage/foo` label and requires one. priority/backlog Higher priority than priority/awaiting-more-evidence.
Projects
Development

No branches or pull requests

9 participants