-
Notifications
You must be signed in to change notification settings - Fork 2.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
kube_pod_status_reason is 0 for all reasons #2612
Comments
Makes sense to mention the version of kube-state-metrics that you used. |
|
Sorry, I'm blind. Didn't correlate the KSM followed by EKS with "kube-state-metrics". |
If it's 0 for all, then https://github.com/kubernetes/kube-state-metrics/blob/main/internal/store/pod.go#L1547 the comparison here might not be correct. |
In the spec for the pod I don't see something like |
/triage accepted |
It's more than a month since it was accepted for the triage. Any updates on this? |
I've pretty much described where setting it to 0 is coming from, feel free to take a look into this and come up with a solution: #2612 (comment) /help |
@mrueg: GuidelinesPlease ensure that the issue body includes answers to the following questions:
For more details on the requirements of such an issue, please see here and ensure that they are met. If this request no longer meets these requirements, the label can be removed In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. |
/assign |
What happened:
The metric
kube_pod_status_reason
shows 0 for all reasons, even when reasons should have value of1
.What you expected to happen:
We use Karpenter in our clusters, and expect to be able to see when pods have a change in status based on actions Karpenter takes. In particular, we expect to see
Evicted
,NodeLost
, andShutdown
reasons to show a value of1
in clusters where consolidation is happening all the time (consolidateAfter
value is5m0s
). We can see in our Karpenter metrics that at any given time, some pod is being moved, and should show up with akube_pod_status_reason
ofEvicted
with a value of1
.How to reproduce it (as minimally and precisely as possible):
This prometheus query:
sum(kube_pod_status_reason) by (reason)
shows0
for every reason, and when charted, those value remain the same over any time interval.Anything else we need to know?:
The
kube_pod_status_phase
does not give use the information we need (specific reasons for status), and no other metric claims to provide this.Environment:
Running KSM v2.13 managed via Helm chart
EKS v1.32.2
Karpenter v1.2.0
The text was updated successfully, but these errors were encountered: