Uploaded image for project: 'OpenShift Bugs'
  1. OpenShift Bugs
  2. OCPBUGS-15928

[Descheduler] - Enabling thresholdPriorityParam in descheduler does not work

XMLWordPrintable

    • Icon: Bug Bug
    • Resolution: Done-Errata
    • Icon: Undefined Undefined
    • None
    • 4.14.0
    • descheduler
    • None
    • Important
    • No
    • Rejected
    • False
    • Hide

      None

      Show
      None

      Description of problem:

      When creating descheduler cluster instance with thresholdPriorityParam i seee that descheduler goes into CrashLoopBackOff state

      Version-Release number of selected component (if applicable):

      4.14

      How reproducible:

      Always

      Steps to Reproduce:

      1. Install 4.14 cluster
      2. Install descheduler operator
      3. Now click on descheduler instance -> try creating instance with thresholdPriorityParam set to prioritym 

      Actual results:

      You could see that descheduler cluster pod goes into crashLoopBackOff state with error belowTolerations:                 node.kubernetes.io/memory-pressure:NoSchedule op=Exists                             node.kubernetes.io/not-ready:NoExecute op=Exists for 300s                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300sEvents:  Type     Reason          Age                 From               Message  ----     ------          ----                ----               -------  Normal   Scheduled       119s                default-scheduler  Successfully assigned openshift-kube-descheduler-operator/descheduler-7f9d68746d-2qr5s to worker-1  Normal   AddedInterface  118s                multus             Add eth0 [10.128.2.65/23] from ovn-kubernetes  Normal   Pulled          29s (x5 over 118s)  kubelet            Container image "registry.redhat.io/openshift4/ose-descheduler@sha256:4bce0bc6b8505a6e6cb13dc2964fb311c17cbefe99e5e6688d646900e97aa641" already present on machine  Normal   Created         29s (x5 over 118s)  kubelet            Created container openshift-descheduler  Normal   Started         29s (x5 over 118s)  kubelet            Started container openshift-descheduler  Warning  BackOff         12s (x9 over 116s)  kubelet            Back-off restarting failed container openshift-descheduler in pod descheduler-7f9d68746d-2qr5s_openshift-kube-descheduler-operator(3842f9f7-616c-4dd3-a441-34a67edaa558)### pod logs output:[root@rdr-varad-ocp-414fl-bastion-0 ~]# oc logs descheduler-7f9d68746d-2qr5sI0706 18:06:14.692859       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/certs-dir/tls.crt::/certs-dir/tls.key"I0706 18:06:15.239729       1 dynamic_serving_content.go:132] "Starting controller" name="serving-cert::/certs-dir/tls.crt::/certs-dir/tls.key"I0706 18:06:15.240270       1 tlsconfig.go:200] "Loaded serving cert" certName="serving-cert::/certs-dir/tls.crt::/certs-dir/tls.key" certDetail="\"metrics.openshift-kube-descheduler-operator.svc\" [serving] validServingFor=[metrics.openshift-kube-descheduler-operator.svc,metrics.openshift-kube-descheduler-operator.svc.cluster.local] issuer=\"openshift-service-serving-signer@1688479304\" (2023-07-06 18:04:33 +0000 UTC to 2025-07-05 18:04:34 +0000 UTC (now=2023-07-06 18:06:15.240223714 +0000 UTC))"I0706 18:06:15.240932       1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1688666775\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1688666775\" (2023-07-06 17:06:14 +0000 UTC to 2024-07-05 17:06:14 +0000 UTC (now=2023-07-06 18:06:15.240893954 +0000 UTC))"I0706 18:06:15.240963       1 secure_serving.go:210] Serving securely on [::]:10258I0706 18:06:15.241058       1 tlsconfig.go:240] "Starting DynamicServingCertificateController"E0706 18:06:15.245389       1 server.go:99] "descheduler server" err="[in profile strategy-LowNodeUtilization-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-PodLifeTime-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-RemoveDuplicates-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-RemovePodsHavingTooManyRestarts-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-RemovePodsViolatingInterPodAntiAffinity-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-RemovePodsViolatingNodeAffinity-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-RemovePodsViolatingNodeTaints-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}, in profile strategy-RemovePodsViolatingTopologySpreadConstraint-profile: priority threshold misconfigured, only one of priorityThreshold fields can be set, got &TypeMeta{Kind:,APIVersion:,}]"I0706 18:06:15.245461       1 tlsconfig.go:255] "Shutting down DynamicServingCertificateController"I0706 18:06:15.245502       1 dynamic_serving_content.go:146] "Shutting down controller" name="serving-cert::/certs-dir/tls.crt::/certs-dir/tls.key"I0706 18:06:15.245638       1 secure_serving.go:255] Stopped listening on [::]:10258Kubedescheduler cluster config:=============================[knarra@knarra openshift-tests-private]$ oc get kubedescheduler cluster -o yaml -n openshift-kube-descheduler-operatorapiVersion: operator.openshift.io/v1kind: KubeDeschedulermetadata:creationTimestamp: "2023-07-06T18:20:17Z"generation: 2name: clusternamespace: openshift-kube-descheduler-operatorresourceVersion: "100884"uid: 57c545ab-106b-49b0-8872-39d59f12bc5bspec:deschedulingIntervalSeconds: 3600logLevel: NormalmanagementState: Managedmode: PredictiveobservedConfig:servingInfo:cipherSuites:- TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256- TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256- TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384- TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384- TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256- TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256minTLSVersion: VersionTLS12operatorLogLevel: NormalprofileCustomizations:devLowNodeUtilizationThresholds: Mediumnamespaces: {}thresholdPriorityClassName: prioritymprofiles:- AffinityAndTaints- TopologyAndDuplicates- LifecycleAndUtilizationunsupportedConfigOverrides: nullstatus:conditions:- lastTransitionTime: "2023-07-06T18:20:18Z"status: "False"type: ConfigObservationDegraded- lastTransitionTime: "2023-07-06T18:20:20Z"status: "False"type: ResourceSyncControllerDegraded- lastTransitionTime: "2023-07-06T18:20:22Z"status: "False"type: TargetConfigControllerDegradedgenerations:- group: appshash: ""lastGeneration: 2name: deschedulernamespace: openshift-kube-descheduler-operatorresource: deploymentsreadyReplicas: 0

      Expected results:

      Descheduler cluster pod should not go into `crashLoopBackOff` state

      Additional info:

      I see the same test works fine in 4.13 cluster.

            rh-ee-lseveroa Lucas Severo Alves
            knarra@redhat.com Rama Kasturi Narra
            Rama Kasturi Narra Rama Kasturi Narra
            Votes:
            0 Vote for this issue
            Watchers:
            4 Start watching this issue

              Created:
              Updated:
              Resolved: