-
Bug
-
Resolution: Not a Bug
-
Major
-
None
-
4.16.z
-
Quality / Stability / Reliability
-
False
-
-
None
-
Important
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
Description of problem:
monitoring cluster operator is degraded and below pods are failing : ~~~ NAME READY STATUS RESTARTS AGE alertmanager-main-0 3/6 CrashLoopBackOff 27 41m alertmanager-main-1 0/6 Init:0/1 0 18m prometheus-k8s-1 0/6 Init:0/1 0 18m ~~~
Version-Release number of selected component (if applicable):
4.16.8
Actual results:
Cluster Operator monitoring is degraded and alertmanager, prometheus pods are failing.
Expected results:
Cluster operator monitoring should be Available and all pods should be in Running state.
Additional info:
alertmanager pod is failing with below error : ~~~ message: | I0207 15:44:45.249649 1 kube-rbac-proxy.go:530] Reading config file: /etc/kube-rbac-proxy/config.yaml E0207 15:44:45.249890 1 run.go:74] "command failed" err="failed to load kubeconfig: cannot find Service Account in pod to build in-cluster rest config: open /var/run/secrets/kubernetes.io/serviceaccount/token: no such file or directory" reason: Error ~~~ Similar BUGs are observed : https://issues.redhat.com/browse/OCPBUGS-30964 , it is fixed in 4.14.z https://issues.redhat.com/browse/OCPBUGS-30954 , it is fixed in 4.16.0 However the customer is running 4.16.8 OCP cluster.