-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is it possible to make the error retrieving resource lock for leader election Fatal? #1280
Comments
The Kubernetes project currently lacks enough contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle rotten |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs. This bot triages issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /close not-planned |
@k8s-triage-robot: Closing this issue, marking it as "Not Planned". In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
E0611 21:13:40.731888 1 leaderelection.go:332] error retrieving resource lock email-operator-system/650681eb.batman.example.com: leases.coordination.k8s.io "650681eb.batman.example.com" is forbidden: User "system:serviceaccount:email-operator-system:email-operator-email-operator-controller-manager" cannot get resource "leases" in API group "coordination.k8s.io" in the namespace "email-operator-system" |
I am running in to this issue too. Three pods were all failing as they were unable to access the lease, and yet the pods remained healthy so the issue went unnoticed. It led to a lot of issues. Happy to raise a PR if we can come up with a solution. I had a look through the code with the hope of finding a non-breaking solution. Exiting with an error could potentially be disruptive to some as current behaviour is during temporary outages there would be a retry. Another option could be adding in a check as part of the
ping to the API.
|
/reopen |
@maggie44: You can't reopen an issue/PR unless you authored it or you are a collaborator. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. |
/reopen |
@smg247: Reopened this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. |
We have recently uncovered an error in our boskos instance like the following:
This went unnoticed for quite some time as it is not a fatal error so the pod wasn't killed (and our alerting was never triggered). Is it possible to make the following errors Fatal?
The text was updated successfully, but these errors were encountered: