Replies: 1 comment 7 replies
-
Anyone who had similar experiences, maybe i´m doing sth wrong, but cluster has around higher spec 4-5 nodes including the control node |
Beta Was this translation helpful? Give feedback.
7 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
How often are your services down?
Did you have the grafana stack installed?
My services are at least down once per month, sometimes clearly because of hetzner and most often because the nodes are cordened
I have to uncorden them manually and it even happens within a new cluster.
I dont know why maybe because of the autoupdate dunno.
The whole experience is quiet frustrating.
Here a summary of error/warnings event within my cluster, which is running since 06/23
Consolidated List of Identified Issues (23.11.23):
Relevant
Volume Attachment and Mounting Issues Attributed to Hetzner:
Node Availability and Scheduling Problems:
System Upgrade Normal Events:
k3s-agent
suggest possible issues causing repeated upgrade attempts.Job Execution Exceeding Deadline in
system-upgrade
Namespace:apply-k3s-agent-on-k3s-watchdog-agent-large-kkf-with-022e-73a05
exceeding its deadline indicates performance issues or misconfiguration.!Relevant
Frequent Normal Events in
trivy-system
Namespace:scan-vulnerabilityreport
objects suggest active scanning processes.Regular Activity in
watchdog
Namespace:prometheus-grafana-stack
andgrafana-stack-kube-prometh-operator
indicate regular monitoring activities.Trivy System Vulnerability Scan Issues:
trivy-system
indicate potential issues with the scanning process.scan-vulnerabilityreport
objects and validate the Trivy configuration.Additional Considerations:
This summary encompasses the issues identified in the provided logs, addressing general Kubernetes cluster concerns, specific volume-related problems, and the activity within the
trivy-system
andwatchdog
namespaces.Beta Was this translation helpful? Give feedback.
All reactions