Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[kube-prometheus-stack] fix(prometheus): rm image!="" from k8s.rules #3863

Closed
wants to merge 2 commits into from

Conversation

jjangga0214
Copy link

@jjangga0214 jjangga0214 commented Oct 5, 2023

What this PR does / why we need it

Which issue this PR fixes

"NO DATA" issues for pod CPU.

BEFORE
스크린샷 2023-10-05 오후 4 06 17

AFTER(PR)
스크린샷 2023-10-05 오후 4 22 10

Special notes for your reviewer

"NO DATA" issues are still happening to others (not cpu usage) metrics (regardless of this PR).
They should be fixed as well, but how?

스크린샷 2023-10-05 오후 4 20 31 스크린샷 2023-10-05 오후 4 21 25

Checklist

  • DCO signed
  • Chart Version bumped
  • Title of the PR starts with chart name (e.g. [prometheus-couchdb-exporter])

@jkroepke
Copy link
Member

@jjangga0214 which CRI are you using? Because omit image!="" would result that CPU usage is twice high.

@jjangga0214
Copy link
Author

jjangga0214 commented Oct 22, 2023

@jkroepke The screenshot was captured against the cluster on "Docker for Desktop" on macOS.
So the CRI is Docker. Maybe this "No Data" issue only happens on "Docker for Desktop"?

@jjangga0214 jjangga0214 changed the title [kube-prometheus-staik] fix(prometheus): rm image!="" from k8s.rules [kube-prometheus-stack] fix(prometheus): rm image!="" from k8s.rules Oct 22, 2023
@jkroepke
Copy link
Member

jkroepke commented Oct 22, 2023

I remember correctly, if CRI is docker. Could be verifiy with minikube where CRI is default to docker and can be configured to containerd.

@jjangga0214
Copy link
Author

jjangga0214 commented Oct 24, 2023

@jkroepke I installed the chart(not modified by this PR) into pure k8s (provisioned by kubespray), but the same issue happens. Its CRI is configured as containerd.

Signed-off-by: Jan-Otto Kröpke <[email protected]>
@jjangga0214
Copy link
Author

jjangga0214 commented Nov 14, 2023

@QuentinBisson Hi! Thanks for the reply.
I don't understand what you are asking for.
Would you please reword more concretely?
Do you mean the content of PrometheusRule CRD from my cluster?

By the way, I just installed v51.10.0 chart, and the issue happens as I described.
I didn't modify the chart.
Then isn't the version itself enough information to provide?

@QuentinBisson
Copy link
Member

@jjangga0214 what I mean is that you should push the change here https://raw.githubusercontent.com/prometheus-operator/kube-prometheus/main/manifests/kubernetesControlPlane-prometheusRule.yaml instead and then come here to sync the rules because we are only consuming the rules from another project

@deefdragon
Copy link

Ive done some thinking and I think the solution hinted at in 3972 (to create a cadvisor chart that gets deployed with the stack) would be a more appropriate path to go down. To deploy a dedicated cadvisor monitor as part of the stack, as opposed to attempting to wrestle with missing metrics would solve the issue for the existing grafana graphs, but also for the Lens K8S UI issues.

I feel having one solution to solve both (and likely more) problems at once is likely better than playing wackamole for each now-broken graph & metric.

Unfortunately, I have no idea what the workload on patching that in would be.

@jkroepke
Copy link
Member

Hi,

I'm closing the PR now. I understand that the error may exists, however as @QuentinBisson mention, the kube-prometheus-stack rules are generated from kube-prometheus and kubernetes-mixin

A manual overwrite would result in a revert. on the next update.

If you still have the issues, consider to create an issue for this.

@jkroepke jkroepke closed this Jan 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants