-
-
Notifications
You must be signed in to change notification settings - Fork 178
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
False-positive "Container has the same readiness and liveness probe" if failureThreshold
differs
#339
Comments
Hey, As I see it, this is by design, to avoid cascading failures, read README_PROBES.md for more information about why this exists. If you don't agree with this, and have examples of why you think that kube-scores recommendations can be harmful, please post them here as a comment, and we can have a chat about it. |
Indeed I'd ask my above production example to be treated as fine. |
Ok, just saw that you submitted a PR that changes the behaviour of The reasoning behind why kube-score tries to make sure that different probes are used, is to make sure that the user understands what they are doing, to avoid making a misconfiguration that leads to unnecessary downtime. Having two probes, with only a difference in either A key difference between a Readiness and Liveness Probe is how the Readiness Probe is used when safely draining connections from a Pod, to make sure that it's unregistered from all Services/Endpoints/LBs/etc before the service is stopped. It's fairly normal to implement a 120s (or more) draining period after receiving SIGTERM, where the Readiness Probe is failing, but the service will still happily keep receiving new requests, until all clients have reacted to the Pod shutting down (by polling the DNS records, or similar). This draining however, can not happen on the Liveness Probe, as restarting the container in the middle of draining connections defeats its purpose. kube-score tries to make it as easy as possible to it's users to spot this type of misconfiguration. Configuring an application to have two different probes, for the two different purposes, is never hard to do, and will make the user spend some time reflecting on their probes, and what the consequences can be it they are misconfigured. And if they can't implement a new probe (if you're running some third party software), this check is easy to be disabled on a per-application basis, with an annotation (but it hopefully still made you think!). I might be wrong, but I'd like to see some examples of where having different time configurations has been the right thing to do, before taking the time to review your PR. Thanks, |
Hello, In fact there is a link in README_PROBES.md to Liveness probes are dangerous, srcco.de. I read here :
|
Yes @greenmaid, that's exactly the main use case. So @zegl any concerns around that? |
The problem with this is there's a single |
Any developments on this? I agree with @AndiDog and would like to "accept" same probe but more lenient in time. Just as the documentation explains it. |
The Kubernetes documentation on probes indicates that "A common pattern for liveness probes is to use the same low-cost HTTP endpoint as for readiness probes, but with a higher failureThreshold". So I would also agree on updating kube-score accordingly. |
In my use case, we have a service that sometimes gets stuck and stops responding to HTTP. I want to use the liveness probe to reboot automatically when this happens. I can't modify the code for legal reasons so I can't change the HTTP path etc. In my case, it makes sense to use the same probe for readiness and liveness, and set the liveness to a higher failure threshold. Sure, it is more clear if I could have different paths/probes. It would also be better if the app didn't get stuck. But I don't think that is related to the actual Kubernetes manifest, which is what kube-score is checking. That's the main reason I feel like this check should be updated. I agree with the commenters - if the liveness probe has a higher failure threshold than readiness, then it should be pretty safe, and it should be allowed by default. I could see some people wanting to opt-in to an even stricter check though, that's fine as long as it is optional. |
This should really be changed since it contradicts the Kubernetes documentation the way it stands today. |
Which version of kube-score are you using?
kube-score version: 1.10.0, commit: 95faa2a, built: 2020-11-07T14:17:50Z
What did you do?
in a Deployment.
What did you expect to see?
No error for this part because
failureThreshold
is a very valid approach to differentiate a liveness probe, for example to kill a container if it's not ready for an extended period of time.What did you see instead?
The text was updated successfully, but these errors were encountered: