-
Bug
-
Resolution: Done-Errata
-
Undefined
-
None
-
4.14.0
-
None
-
Important
-
No
-
Rejected
-
False
-
Description of problem:
When creating descheduler cluster instance with thresholdPriorityParam i seee that descheduler goes into CrashLoopBackOff state
Version-Release number of selected component (if applicable):
4.14
How reproducible:
Always
Steps to Reproduce:
1. Install 4.14 cluster 2. Install descheduler operator 3. Now click on descheduler instance -> try creating instance with thresholdPriorityParam set to prioritym
Actual results:
You could see that descheduler cluster pod goes into crashLoopBackOff state with error belowTolerations: node.kubernetes.io/memory-pressure:NoSchedule op=Exists node.kubernetes.io/not-ready:NoExecute op=Exists for 300s node.kubernetes.io/unreachable:NoExecute op=Exists for 300sEvents: Type Reason Age From Message ---- ------ ---- ---- ------- Normal Scheduled 119s default-scheduler Successfully assigned openshift-kube-descheduler-operator/descheduler-7f9d68746d-2qr5s to worker-1 Normal AddedInterface 118s multus Add eth0 [10.128.2.65/23] from ovn-kubernetes Normal Pulled 29s (x5 over 118s) kubelet Container image "registry.redhat.io/openshift4/ose-descheduler@sha256:4bce0bc6b8505a6e6cb13dc2964fb311c17cbefe99e5e6688d646900e97aa641" already present on machine Normal Created 29s (x5 over 118s) kubelet Created container openshift-descheduler Normal Started 29s (x5 over 118s) kubelet Started container openshift-descheduler Warning BackOff 12s (x9 over 116s) kubelet Back-off restarting failed container openshift-descheduler in pod descheduler-7f9d68746d-2qr5s_openshift-kube-descheduler-operator(3842f9f7-616c-4dd3-a441-34a67edaa558)### pod logs output:[root@rdr-varad-ocp-414fl-bastion-0 ~]# oc logs descheduler-7f9d68746d-2qr5sI0706 18:06:14.692859 1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/certs-dir/tls.crt::/certs-dir/tls.key"I0706 18:06:15.239729 1 dynamic_serving_content.go:132] "Starting controller" name="serving-cert::/certs-dir/tls.crt::/certs-dir/tls.key"I0706 18:06:15.240270 1 tlsconfig.go:200] "Loaded serving cert" certName="serving-cert::/certs-dir/tls.crt::/certs-dir/tls.key" certDetail="\"metrics.openshift-kube-descheduler-operator.svc\" [serving] validServingFor=[metrics.openshift-kube-descheduler-operator.svc,metrics.openshift-kube-descheduler-operator.svc.cluster.local] issuer=\"openshift-service-serving-signer@1688479304\" (2023-07-06 18:04:33 +0000 UTC to 2025-07-05 18:04:34 +0000 UTC (now=2023-07-06 18:06:15.240223714 +0000 UTC))"I0706 18:06:15.240932 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1688666775\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1688666775\" (2023-07-06 17:06:14 +0000 UTC to 2024-07-05 17:06:14 +0000 UTC (now=2023-07-06 18:06:15.240893954 +0000 UTC))"I0706 18:06:15.240963 1 secure_serving.go:210] Serving securely on [::]:10258I0706 18:06:15.241058 1 tlsconfig.go:240] "Starting DynamicServingCertificateController"E0706 18:06:15.245389 1 server.go:99] "descheduler server" err="[in profile strategy-LowNodeUtilization-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-PodLifeTime-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-RemoveDuplicates-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-RemovePodsHavingTooManyRestarts-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-RemovePodsViolatingInterPodAntiAffinity-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-RemovePodsViolatingNodeAffinity-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-RemovePodsViolatingNodeTaints-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-RemovePodsViolatingTopologySpreadConstraint-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}]"I0706 18:06:15.245461 1 tlsconfig.go:255] "Shutting down DynamicServingCertificateController"I0706 18:06:15.245502 1 dynamic_serving_content.go:146] "Shutting down controller" name="serving-cert::/certs-dir/tls.crt::/certs-dir/tls.key"I0706 18:06:15.245638 1 secure_serving.go:255] Stopped listening on [::]:10258Kubedescheduler cluster config:=============================[knarra@knarra openshift-tests-private]$ oc get kubedescheduler cluster -o yaml -n openshift-kube-descheduler-operatorapiVersion: operator.openshift.io/v1kind: KubeDeschedulermetadata:creationTimestamp: "2023-07-06T18:20:17Z"generation: 2name: clusternamespace: openshift-kube-descheduler-operatorresourceVersion: "100884"uid: 57c545ab-106b-49b0-8872-39d59f12bc5bspec:deschedulingIntervalSeconds: 3600logLevel: NormalmanagementState: Managedmode: PredictiveobservedConfig:servingInfo:cipherSuites:- TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256- TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256- TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384- TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384- TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256- TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256minTLSVersion: VersionTLS12operatorLogLevel: NormalprofileCustomizations:devLowNodeUtilizationThresholds: Mediumnamespaces: {}thresholdPriorityClassName: prioritymprofiles:- AffinityAndTaints- TopologyAndDuplicates- LifecycleAndUtilizationunsupportedConfigOverrides: nullstatus:conditions:- lastTransitionTime: "2023-07-06T18:20:18Z"status: "False"type: ConfigObservationDegraded- lastTransitionTime: "2023-07-06T18:20:20Z"status: "False"type: ResourceSyncControllerDegraded- lastTransitionTime: "2023-07-06T18:20:22Z"status: "False"type: TargetConfigControllerDegradedgenerations:- group: appshash: ""lastGeneration: 2name: deschedulernamespace: openshift-kube-descheduler-operatorresource: deploymentsreadyReplicas: 0
Expected results:
Descheduler cluster pod should not go into `crashLoopBackOff` state
Additional info:
I see the same test works fine in 4.13 cluster.