Uploaded image for project: 'Network Observability'
  1. Network Observability
  2. NETOBSERV-983

eBPF pod requires more memory than the default limit

    • Quality / Stability / Reliability
    • False
    • Hide

      None

      Show
      None
    • None
    • Low
    • None
    • NetObserv - Sprint 236, NetObserv - Sprint 237, NetObserv - Sprint 238, NetObserv - Sprint 239
    • None
    • None
    • None

      We've observed over several rounds of testing NetObserv on our Baremetal cluster where one or more eBPF pods with be OOMKilled and go into CrashLoopBackOff state: https://docs.google.com/document/d/1DOfV17DEuqI0YSW6oOLc_XQlze-ZFS5FtqYf39n179E/edit?usp=sharing

      This can be mitigated by increasing the eBPF memory limit from the default 800Mi - we've seen success with a limit of 2000Mi though this is likely tied to the amount of traffic we are generating.

      Marking this as low severity as it doesn't seem to result in dropped flows, but we should either increase the default memory limit for eBPF or try and find a way to balance this load across other eBPF pods.

              mmahmoud@redhat.com Mohamed Mahmoud (Inactive)
              nweinber1 Nathan Weinberg
              None
              None
              None
              Nathan Weinberg Nathan Weinberg
              None
              Votes:
              0 Vote for this issue
              Watchers:
              3 Start watching this issue

                Created:
                Updated:
                Resolved: