Uploaded image for project: 'OpenShift Bugs'
  1. OpenShift Bugs
  2. OCPBUGS-29713

Excessive node status updates causing high control plane CPU

XMLWordPrintable

    • +
    • No
    • MCO Sprint 249, MCO Sprint 250
    • 2
    • Rejected
    • False
    • Hide

      None

      Show
      None
    • Hide
      When nodeStatusUpdateFrequency's default changed from 0s to 10s in the MCO, we inadvertently caused the nodeStatusReportFrequency to increase as the logic to set its higher value was linked to the 0s of the nodeStatusReportFrequency. We need to manually set nodeStatusReportFrequency to 5m
      Show
      When nodeStatusUpdateFrequency's default changed from 0s to 10s in the MCO, we inadvertently caused the nodeStatusReportFrequency to increase as the logic to set its higher value was linked to the 0s of the nodeStatusReportFrequency. We need to manually set nodeStatusReportFrequency to 5m
    • Bug Fix
    • In Progress

      Description of problem:

      OCPBUGS-29424 revealed that setting the node status update frequency in kubelet (introduced with OCPBUGS-15583) causes a lot of control plane CPU. 
      
      The reason is the increased frequency of kubelet node status updates will trigger second order effects in all control plane operators that usually trigger on node changes (api server, etcd, PDB guard pod controllers, or any other static pod based machinery).
      
      Reverting the code in OCPBUGS-15583, or manually setting the report/status frequency to 0s causes the CPU to drop immediately. 

      Version-Release number of selected component (if applicable):

      Versions where OCPBUGS-15583 was backported. This includes 4.16, 4.15.0, 4.14.8, 4.13.33, and the next 4.12.z likely 4.12.51.

      How reproducible:

      always    

      Steps to Reproduce:

      1. create a cluster that contains a fix for OCPBUGS-15583
      2. observe the apiserver metrics (eg rate(apiserver_request_total[5m])), those should show abnormal values for pod/configmap GET
          alternatively the rate of node updates is increaed (rate(apiserver_request_total{resource="nodes", subresource="status", verb="PATCH"}[1m])) 
      
           

      Actual results:

      the node status updates every 10s, which causes high CPU usage on control plane operators and apiserver

      Expected results:

      the node status should not update that frequently, meaning the control plane CPU usage should go down again 

      Additional info:

      slack thread with the node team:
      https://redhat-internal.slack.com/archives/C02CZNQHGN8/p1708429189987849
          

            cdoern@redhat.com Charles Doern
            tjungblu@redhat.com Thomas Jungblut
            Sergio Regidor de la Rosa Sergio Regidor de la Rosa
            Votes:
            0 Vote for this issue
            Watchers:
            11 Start watching this issue

              Created:
              Updated: