-
Bug
-
Resolution: Done
-
Critical
-
None
-
4.12.0
-
None
-
Quality / Stability / Reliability
-
False
-
-
None
-
None
-
None
-
None
-
Rejected
-
SDN Sprint 226, SDN Sprint 227
-
2
-
None
-
None
-
None
-
None
-
None
-
None
-
None
Description of problem:
The daemon pods running perfectly starts restarting on the node that has sctp module loaded on the node.
Version-Release number of selected component (if applicable):
4.12
How reproducible:
Always
Steps to Reproduce:
1. Install the ingress node firewall operator and create a ingress node firewall config.
oc get csv -n openshift-ingress-node-firewall
NAME DISPLAY VERSION REPLACES PHASE
ingress-node-firewall.4.12.0-202210262313 ingress-node-firewall-operator 4.12.0-202210262313 Succeeded oc get pods -n openshift-ingress-node-firewall -owide
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
ingress-node-firewall-controller-manager-d6cb6c859-n2979 2/2 Running 0 7m57s 10.128.2.10 asood-10273-w6xnn-worker-mqzj7 <none> <none>
ingress-node-firewall-daemon-bwx6t 3/3 Running 0 4m27s 172.31.249.38 asood-10273-w6xnn-worker-h5l2k <none> <none>
ingress-node-firewall-daemon-d45j5 3/3 Running 0 4m27s 172.31.249.163 asood-10273-w6xnn-worker-mqzj7 <none> <none>
2. Load the sctp module on the worker nodes and wait for the nodes to be return to ready state.
--
apiVersion: machineconfiguration.openshift.io/v1
kind: MachineConfig
metadata:
labels:
machineconfiguration.openshift.io/role: worker
name: load-sctp-module
spec:
config:
ignition:
version: 2.2.0
storage:
files:
- contents:
source: data:,
verification: {}
filesystem: root
mode: 420
path: /etc/modprobe.d/sctp-blacklist.conf
- contents:
source: data:text/plain;charset=utf-8,sctp
filesystem: root
mode: 420
path: /etc/modules-load.d/sctp-load.conf
[asood@asood ~]$ oc get nodes
NAME STATUS ROLES AGE VERSION
asood-10273-w6xnn-master-0 Ready control-plane,master 75m v1.25.2+4bd0702
asood-10273-w6xnn-master-1 Ready control-plane,master 75m v1.25.2+4bd0702
asood-10273-w6xnn-master-2 Ready control-plane,master 75m v1.25.2+4bd0702
asood-10273-w6xnn-worker-h5l2k NotReady,SchedulingDisabled worker 44m v1.25.2+4bd0702
asood-10273-w6xnn-worker-mqzj7 Ready worker 44m v1.25.2+4bd0702
[asood@asood ~]$ oc get nodes
NAME STATUS ROLES AGE VERSION
asood-10273-w6xnn-master-0 Ready control-plane,master 76m v1.25.2+4bd0702
asood-10273-w6xnn-master-1 Ready control-plane,master 76m v1.25.2+4bd0702
asood-10273-w6xnn-master-2 Ready control-plane,master 76m v1.25.2+4bd0702
asood-10273-w6xnn-worker-h5l2k Ready worker 46m v1.25.2+4bd0702
asood-10273-w6xnn-worker-mqzj7 Ready,SchedulingDisabled worker 46m v1.25.2+4bd0702
[asood@asood ~]$ oc get nodes
NAME STATUS ROLES AGE VERSION
asood-10273-w6xnn-master-0 Ready control-plane,master 84m v1.25.2+4bd0702
asood-10273-w6xnn-master-1 Ready control-plane,master 84m v1.25.2+4bd0702
asood-10273-w6xnn-master-2 Ready control-plane,master 84m v1.25.2+4bd0702
asood-10273-w6xnn-worker-h5l2k Ready worker 54m v1.25.2+4bd0702
asood-10273-w6xnn-worker-mqzj7 Ready worker 54m v1.25.2+4bd0702
[asood@asood ~]$ oc get pods -n openshift-ingress-node-firewall -owide
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
ingress-node-firewall-controller-manager-d6cb6c859-sv9vn 2/2 Running 0 27s 10.131.0.23 asood-10273-w6xnn-worker-h5l2k <none> <none>
ingress-node-firewall-daemon-bwx6t 2/3 Error 6 8m 172.31.249.38 asood-10273-w6xnn-worker-h5l2k <none> <none>
ingress-node-firewall-daemon-d45j5 3/3 Running 0 8m 172.31.249.163 asood-10273-w6xnn-worker-mqzj7 <none> <none>
[asood@asood ~]$ oc get pods -n openshift-ingress-node-firewall -owide
NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES
ingress-node-firewall-controller-manager-d6cb6c859-sv9vn 2/2 Running 0 8m58s 10.131.0.23 asood-10273-w6xnn-worker-h5l2k <none> <none>
ingress-node-firewall-daemon-bwx6t 2/3 CrashLoopBackOff 9 (3m33s ago) 16m 172.31.249.38 asood-10273-w6xnn-worker-h5l2k <none> <none>
ingress-node-firewall-daemon-d45j5 2/3 CrashLoopBackOff 8 (17s ago) 16m 172.31.249.163 asood-10273-w6xnn-worker-mqzj7 <none> <none>
Actual results:
The ingress node firewall pods end up in crashloop state
Expected results:
The pods should continue to be in ready state.
Additional info:
The cluster is installed on vSphere. ipi-on-vsphere/versioned-installer-vmc7-ovn
oc debug node/asood-10273-w6xnn-worker-h5l2k
Starting pod/asood-10273-w6xnn-worker-h5l2k-debug ...
To use host binaries, run `chroot /host`
Pod IP: 172.31.249.38
If you don't see a command prompt, try pressing enter.
sh-4.4# chroot /host
sh-4.4# lsmod | grep sctp
sh-4.4# exit
exit
sh-4.4# exit
exit
After loading SCTP module.
oc debug node/asood-10273-w6xnn-worker-h5l2k
Starting pod/asood-10273-w6xnn-worker-h5l2k-debug ...
To use host binaries, run `chroot /host`
Pod IP: 172.31.249.38
If you don't see a command prompt, try pressing enter.
sh-4.4# chroot /host
sh-4.4# lsmod | grep sctp
sctp 421888 34
ip6_udp_tunnel 16384 2 geneve,sctp
udp_tunnel 20480 2 geneve,sctp
libcrc32c 16384 6 nf_conntrack,nf_nat,openvswitch,nf_tables,xfs,sctp
sh-4.4#
oc describe pod ingress-node-firewall-daemon-bwx6t -n openshift-ingress-node-firewall
Name: ingress-node-firewall-daemon-bwx6t
Namespace: openshift-ingress-node-firewall
Priority: 2000001000
Priority Class Name: system-node-critical
Node: asood-10273-w6xnn-worker-h5l2k/172.31.249.38
Start Time: Thu, 27 Oct 2022 17:52:12 -0400
Labels: app=ingress-node-firewall-daemon
component=daemon
controller-revision-hash=f7b68c595
pod-template-generation=1
type=infra
Annotations: openshift.io/scc: privileged
Status: Running
IP: 172.31.249.38
IPs:
IP: 172.31.249.38
Controlled By: DaemonSet/ingress-node-firewall-daemon
Containers:
daemon:
Container ID: cri-o://92896263dfba382986cedf2ae300632c0a55e97c41c90d6b9004ebfafc1112a3
Image: registry.redhat.io/openshift4/ingress-node-firewall-daemon@sha256:a7c7337ffcb9e17608a8d8bf701af9ad3e0c11e3a9b93fb912558cee1208c4f9
Image ID: registry.redhat.io/openshift4/ingress-node-firewall-daemon@sha256:a7c7337ffcb9e17608a8d8bf701af9ad3e0c11e3a9b93fb912558cee1208c4f9
Port: <none>
Host Port: <none>
State: Running
Started: Thu, 27 Oct 2022 17:59:25 -0400
Ready: True
Restart Count: 1
Environment:
NODE_NAME: (v1:spec.nodeName)
NAMESPACE: openshift-ingress-node-firewall (v1:metadata.namespace)
POLL_PERIOD_SECONDS: 30
Mounts:
/sys/fs/bpf from bpf-maps (rw)
/var/run from syslog-socket (rw)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-4qczm (ro)
events:
Container ID: cri-o://2686e8168b13d680e6358c378f6091874f6369f6028fd516d41b57d365e7e03f
Image: registry.redhat.io/openshift4/ingress-node-firewall-daemon@sha256:a7c7337ffcb9e17608a8d8bf701af9ad3e0c11e3a9b93fb912558cee1208c4f9
Image ID: registry.redhat.io/openshift4/ingress-node-firewall-daemon@sha256:a7c7337ffcb9e17608a8d8bf701af9ad3e0c11e3a9b93fb912558cee1208c4f9
Port: <none>
Host Port: <none>
Command:
/usr/bin/syslog
State: Waiting
Reason: CrashLoopBackOff
Last State: Terminated
Reason: Error
Exit Code: 1
Started: Thu, 27 Oct 2022 18:10:17 -0400
Finished: Thu, 27 Oct 2022 18:10:17 -0400
Ready: False
Restart Count: 8
Requests:
cpu: 100m
memory: 256Mi
Environment: <none>
Mounts:
/var/run from syslog-socket (rw)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-4qczm (ro)
kube-rbac-proxy:
Container ID: cri-o://15b6df8abeb9595695200f10341e2cb8ccd6861497ef054a06e537cc771d1927
Image: registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ed81d739e83332a72459fe6b289b490bf53c3ca97d5af9fb34dbe98f7e99c6f
Image ID: quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ed81d739e83332a72459fe6b289b490bf53c3ca97d5af9fb34dbe98f7e99c6f
Port: 9301/TCP
Host Port: 9301/TCP
Command:
/bin/bash
-c
#!/bin/bash
set -euo pipefail
TLS_PK=/etc/pki/tls/metrics-certs/tls.key
TLS_CERT=/etc/pki/tls/metrics-certs/tls.crt
# As the secret mount is optional we must wait for the files to be present.
# If it isn't created there is probably an issue so we want to crashloop.
TS=$(date +%s)
WARN_TS=$(( ${TS} + $(( 20 * 60)) ))
HAS_LOGGED_INFO=0
log_missing_certs(){
CUR_TS=$(date +%s)
if [[ "${CUR_TS}" -gt "${WARN_TS}" ]]; then
echo $(date -Iseconds) WARN: ingress-node-firewall-daemon-metrics-certs not mounted after 20 minutes.
elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then
echo $(date -Iseconds) INFO: ingress-node-firewall-daemon-metrics-certs not mounted. Waiting 20 minutes.
HAS_LOGGED_INFO=1
fi
}
while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do
log_missing_certs
sleep 5
done
echo $(date -Iseconds) INFO: ingress-node-firewall-daemon-metrics-certs mounted, starting kube-rbac-proxy
exec /usr/bin/kube-rbac-proxy \
--logtostderr \
--secure-listen-address=:9301 \
--tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 \
--upstream=http://127.0.0.1:39301 / \
--tls-private-key-file=${TLS_PK} \
--tls-cert-file=${TLS_CERT}
State: Running
Started: Thu, 27 Oct 2022 17:59:26 -0400
Ready: True
Restart Count: 1
Requests:
cpu: 10m
memory: 20Mi
Environment: <none>
Mounts:
/etc/pki/tls/metrics-certs from ingress-node-firewall-daemon-metrics-certs (ro)
/var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-4qczm (ro)
Conditions:
Type Status
Initialized True
Ready False
ContainersReady False
PodScheduled True
Volumes:
bpf-maps:
Type: HostPath (bare host directory volume)
Path: /sys/fs/bpf
HostPathType: DirectoryOrCreate
ingress-node-firewall-daemon-metrics-certs:
Type: Secret (a volume populated by a Secret)
SecretName: ingress-node-firewall-daemon-metrics-certs
Optional: true
syslog-socket:
Type: EmptyDir (a temporary directory that shares a pod's lifetime)
Medium:
SizeLimit: <unset>
kube-api-access-4qczm:
Type: Projected (a volume that contains injected data from multiple sources)
TokenExpirationSeconds: 3607
ConfigMapName: kube-root-ca.crt
ConfigMapOptional: <nil>
DownwardAPI: true
ConfigMapName: openshift-service-ca.crt
ConfigMapOptional: <nil>
QoS Class: Burstable
Node-Selectors: node-role.kubernetes.io/worker=
Tolerations: op=Exists
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 20m default-scheduler Successfully assigned openshift-ingress-node-firewall/ingress-node-firewall-daemon-bwx6t to asood-10273-w6xnn-worker-h5l2k
Normal Pulling 20m kubelet Pulling image "registry.redhat.io/openshift4/ingress-node-firewall-daemon@sha256:a7c7337ffcb9e17608a8d8bf701af9ad3e0c11e3a9b93fb912558cee1208c4f9"
Normal Pulled 20m kubelet Successfully pulled image "registry.redhat.io/openshift4/ingress-node-firewall-daemon@sha256:a7c7337ffcb9e17608a8d8bf701af9ad3e0c11e3a9b93fb912558cee1208c4f9" in 3.678735249s
Normal Created 20m kubelet Created container daemon
Normal Started 20m kubelet Started container daemon
Normal Pulled 20m kubelet Container image "registry.redhat.io/openshift4/ingress-node-firewall-daemon@sha256:a7c7337ffcb9e17608a8d8bf701af9ad3e0c11e3a9b93fb912558cee1208c4f9" already present on machine
Normal Created 20m kubelet Created container events
Normal Started 20m kubelet Started container events
Normal Pulling 20m kubelet Pulling image "registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ed81d739e83332a72459fe6b289b490bf53c3ca97d5af9fb34dbe98f7e99c6f"
Normal Pulled 20m kubelet Successfully pulled image "registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ed81d739e83332a72459fe6b289b490bf53c3ca97d5af9fb34dbe98f7e99c6f" in 2.602971686s
Normal Created 20m kubelet Created container kube-rbac-proxy
Normal Started 20m kubelet Started container kube-rbac-proxy
Warning NodeNotReady 17m node-controller Node is not ready
Normal Pulled 13m kubelet Container image "registry.redhat.io/openshift4/ingress-node-firewall-daemon@sha256:a7c7337ffcb9e17608a8d8bf701af9ad3e0c11e3a9b93fb912558cee1208c4f9" already present on machine
Normal Created 13m kubelet Created container daemon
Normal Started 13m kubelet Started container daemon
Normal Pulled 13m kubelet Container image "registry.redhat.io/openshift4/ose-kube-rbac-proxy@sha256:6ed81d739e83332a72459fe6b289b490bf53c3ca97d5af9fb34dbe98f7e99c6f" already present on machine
Normal Created 13m kubelet Created container kube-rbac-proxy
Normal Started 13m kubelet Started container kube-rbac-proxy
Normal Pulled 12m (x4 over 13m) kubelet Container image "registry.redhat.io/openshift4/ingress-node-firewall-daemon@sha256:a7c7337ffcb9e17608a8d8bf701af9ad3e0c11e3a9b93fb912558cee1208c4f9" already present on machine
Normal Created 12m (x4 over 13m) kubelet Created container events
Normal Started 12m (x4 over 13m) kubelet Started container events
Warning BackOff 3m9s (x49 over 13m) kubelet Back-off restarting failed container
oc logs ingress-node-firewall-controller-manager-d6cb6c859-sv9vn -n openshift-ingress-node-firewall
1.6669079899095576e+09 INFO setup Version {"version.Version": "4.12.0"}
I1027 21:59:50.960910 1 request.go:682] Waited for 1.037223877s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/apis/console.openshift.io/v1?timeout=32s
1.6669079922668066e+09 INFO controller-runtime.metrics Metrics server is starting to listen {"addr": "127.0.0.1:39300"}
1.666907992268518e+09 INFO controller-runtime.builder skip registering a mutating webhook, object does not implement admission.Defaulter or WithDefaulter wasn't called {"GVK": "ingressnodefirewall.openshift.io/v1alpha1, Kind=IngressNodeFirewall"}
1.6669079922685819e+09 INFO controller-runtime.builder Registering a validating webhook {"GVK": "ingressnodefirewall.openshift.io/v1alpha1, Kind=IngressNodeFirewall", "path": "/validate-ingressnodefirewall-openshift-io-v1alpha1-ingressnodefirewall"}
1.666907992268735e+09 INFO controller-runtime.webhook Registering webhook {"path": "/validate-ingressnodefirewall-openshift-io-v1alpha1-ingressnodefirewall"}
1.6669079922691386e+09 INFO platform detecting platform version...
1.666907992275584e+09 INFO platform route.openshift.io found in apis, platform is OpenShift
1.6669079922756834e+09 INFO platform PlatformInfo [Name: OpenShift, K8SVersion: 1.25, OS: linux/amd64]
1.6669079922757583e+09 INFO setup starting manager
1.666907992276205e+09 INFO controller-runtime.webhook.webhooks Starting webhook server
1.666907992276267e+09 INFO Starting server {"kind": "health probe", "addr": "[::]:8081"}
1.6669079922762952e+09 INFO Starting server {"path": "/metrics", "kind": "metrics", "addr": "127.0.0.1:39300"}
I1027 21:59:52.276877 1 leaderelection.go:248] attempting to acquire leader lease openshift-ingress-node-firewall/d902e78d.ingress-nodefw...
1.6669079922770069e+09 INFO controller-runtime.certwatcher Updated current TLS certificate
1.6669079922771842e+09 INFO controller-runtime.webhook Serving webhook server {"host": "", "port": 9443}
1.666907992277203e+09 INFO controller-runtime.certwatcher Starting certificate watcher
I1027 22:00:10.113724 1 leaderelection.go:258] successfully acquired lease openshift-ingress-node-firewall/d902e78d.ingress-nodefw
1.6669080101138086e+09 DEBUG events ingress-node-firewall-controller-manager-d6cb6c859-sv9vn_26498be8-e584-4131-8b0b-290c7a990f6b became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"openshift-ingress-node-firewall","name":"d902e78d.ingress-nodefw","uid":"18f2c4e6-ddf3-4644-9f70-88298a8db49d","apiVersion":"coordination.k8s.io/v1","resourceVersion":"60958"}, "reason": "LeaderElection"}
1.666908010113973e+09 INFO Starting EventSource {"controller": "ingressnodefirewall", "controllerGroup": "ingressnodefirewall.openshift.io", "controllerKind": "IngressNodeFirewall", "source": "kind source: *v1alpha1.IngressNodeFirewall"}
1.6669080101140242e+09 INFO Starting EventSource {"controller": "ingressnodefirewall", "controllerGroup": "ingressnodefirewall.openshift.io", "controllerKind": "IngressNodeFirewall", "source": "kind source: *v1.Node"}
1.6669080101140344e+09 INFO Starting EventSource {"controller": "ingressnodefirewall", "controllerGroup": "ingressnodefirewall.openshift.io", "controllerKind": "IngressNodeFirewall", "source": "kind source: *v1alpha1.IngressNodeFirewallNodeState"}
1.66690801011404e+09 INFO Starting Controller {"controller": "ingressnodefirewall", "controllerGroup": "ingressnodefirewall.openshift.io", "controllerKind": "IngressNodeFirewall"}
1.666908010114034e+09 INFO Starting EventSource {"controller": "ingressnodefirewallconfig", "controllerGroup": "ingressnodefirewall.openshift.io", "controllerKind": "IngressNodeFirewallConfig", "source": "kind source: *v1alpha1.IngressNodeFirewallConfig"}
1.6669080101140623e+09 INFO Starting EventSource {"controller": "ingressnodefirewallconfig", "controllerGroup": "ingressnodefirewall.openshift.io", "controllerKind": "IngressNodeFirewallConfig", "source": "kind source: *v1.DaemonSet"}
1.6669080101140668e+09 INFO Starting Controller {"controller": "ingressnodefirewallconfig", "controllerGroup": "ingressnodefirewall.openshift.io", "controllerKind": "IngressNodeFirewallConfig"}
1.6669080102193794e+09 INFO Starting workers {"controller": "ingressnodefirewall", "controllerGroup": "ingressnodefirewall.openshift.io", "controllerKind": "IngressNodeFirewall", "worker count": 1}
1.6669080102194178e+09 INFO Starting workers {"controller": "ingressnodefirewallconfig", "controllerGroup": "ingressnodefirewall.openshift.io", "controllerKind": "IngressNodeFirewallConfig", "worker count": 1}
1.6669080102196178e+09 INFO controllers.IngressNodeFirewallConfig.syncIngressNodeFirewallConfigResources Start
2022/10/27 22:00:10 reconciling (apps/v1, Kind=DaemonSet) openshift-ingress-node-firewall/ingress-node-firewall-daemon
2022/10/27 22:00:10 update was successful
1.666908010260744e+09 INFO controllers.IngressNodeFirewallConfig.syncIngressNodeFirewallConfigResources Start
2022/10/27 22:00:10 reconciling (apps/v1, Kind=DaemonSet) openshift-ingress-node-firewall/ingress-node-firewall-daemon
2022/10/27 22:00:10 update was successful
1.6669080103077493e+09 ERROR controllers.IngressNodeFirewallConfig Failed to update ingress node firewall config status {"ingress node firewall config": "openshift-ingress-node-firewall/ingressnodefirewallconfig", "Desired status": "Available", "error": "could not update status for object &{TypeMeta:{Kind:IngressNodeFirewallConfig APIVersion:ingressnodefirewall.openshift.io/v1alpha1} ObjectMeta:{Name:ingressnodefirewallconfig GenerateName: Namespace:openshift-ingress-node-firewall SelfLink: UID:a31e4bc5-2262-41af-8262-8c0934be19e1 ResourceVersion:57963 Generation:1 CreationTimestamp:2022-10-27 21:52:12 +0000 UTC DeletionTimestamp:<nil> DeletionGracePeriodSeconds:<nil> Labels:map[] Annotations:map[] OwnerReferences:[] Finalizers:[] ManagedFields:[{Manager:kubectl-create Operation:Update APIVersion:ingressnodefirewall.openshift.io/v1alpha1 Time:2022-10-27 21:52:12 +0000 UTC FieldsType:FieldsV1 FieldsV1:{\"f:spec\":{\".\":{},\"f:nodeSelector\":{\".\":{},\"f:node-role.kubernetes.io/worker\":{}}}} Subresource:} {Manager:manager Operation:Update APIVersion:ingressnodefirewall.openshift.io/v1alpha1 Time:2022-10-27 21:55:44 +0000 UTC FieldsType:FieldsV1 FieldsV1:{\"f:status\":{\".\":{},\"f:conditions\":{}}} Subresource:status}]} Spec:{NodeSelector:map[node-role.kubernetes.io/worker:]} Status:{Conditions:[{Type:Available Status:True ObservedGeneration:0 LastTransitionTime:2022-10-27 22:00:10.290982302 +0000 UTC m=+20.432947581 Reason:Available Message:} {Type:Progressing Status:False ObservedGeneration:0 LastTransitionTime:2022-10-27 22:00:10.290982302 +0000 UTC m=+20.432947581 Reason:Progressing Message:} {Type:Degraded Status:False ObservedGeneration:0 LastTransitionTime:2022-10-27 22:00:10.290982302 +0000 UTC m=+20.432947581 Reason:Degraded Message:}]}}: Operation cannot be fulfilled on ingressnodefirewallconfigs.ingressnodefirewall.openshift.io \"ingressnodefirewallconfig\": the object has been modified; please apply your changes to the latest version and try again", "errorVerbose": "Operation cannot be fulfilled on ingressnodefirewallconfigs.ingressnodefirewall.openshift.io \"ingressnodefirewallconfig\": the object has been modified; please apply your changes to the latest version and try again\ncould not update status for object &{TypeMeta:{Kind:IngressNodeFirewallConfig APIVersion:ingressnodefirewall.openshift.io/v1alpha1} ObjectMeta:{Name:ingressnodefirewallconfig GenerateName: Namespace:openshift-ingress-node-firewall SelfLink: UID:a31e4bc5-2262-41af-8262-8c0934be19e1 ResourceVersion:57963 Generation:1 CreationTimestamp:2022-10-27 21:52:12 +0000 UTC DeletionTimestamp:<nil> DeletionGracePeriodSeconds:<nil> Labels:map[] Annotations:map[] OwnerReferences:[] Finalizers:[] ManagedFields:[{Manager:kubectl-create Operation:Update APIVersion:ingressnodefirewall.openshift.io/v1alpha1 Time:2022-10-27 21:52:12 +0000 UTC FieldsType:FieldsV1 FieldsV1:{\"f:spec\":{\".\":{},\"f:nodeSelector\":{\".\":{},\"f:node-role.kubernetes.io/worker\":{}}}} Subresource:} {Manager:manager Operation:Update APIVersion:ingressnodefirewall.openshift.io/v1alpha1 Time:2022-10-27 21:55:44 +0000 UTC FieldsType:FieldsV1 FieldsV1:{\"f:status\":{\".\":{},\"f:conditions\":{}}} Subresource:status}]} Spec:{NodeSelector:map[node-role.kubernetes.io/worker:]} Status:{Conditions:[{Type:Available Status:True ObservedGeneration:0 LastTransitionTime:2022-10-27 22:00:10.290982302 +0000 UTC m=+20.432947581 Reason:Available Message:} {Type:Progressing Status:False ObservedGeneration:0 LastTransitionTime:2022-10-27 22:00:10.290982302 +0000 UTC m=+20.432947581 Reason:Progressing Message:} {Type:Degraded Status:False ObservedGeneration:0 LastTransitionTime:2022-10-27 22:00:10.290982302 +0000 UTC m=+20.432947581 Reason:Degraded Message:}]}}\ngithub.com/openshift/ingress-node-firewall/pkg/status.Update\n\t/workspace/pkg/status/status.go:49\ngithub.com/openshift/ingress-node-firewall/controllers.(*IngressNodeFirewallConfigReconciler).Reconcile\n\t/workspace/controllers/ingressnodefirewallconfig_controller.go:115\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:121\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:320\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:273\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:234\nruntime.goexit\n\t/usr/lib/golang/src/runtime/asm_amd64.s:1594"}
github.com/openshift/ingress-node-firewall/controllers.(*IngressNodeFirewallConfigReconciler).Reconcile
/workspace/controllers/ingressnodefirewallconfig_controller.go:116
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile
/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:121
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler
/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:320
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem
/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:273
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2
/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:234
1.666908010307969e+09 ERROR Reconciler error {"controller": "ingressnodefirewallconfig", "controllerGroup": "ingressnodefirewall.openshift.io", "controllerKind": "IngressNodeFirewallConfig", "IngressNodeFirewallConfig": {"name":"ingressnodefirewallconfig","namespace":"openshift-ingress-node-firewall"}, "namespace": "openshift-ingress-node-firewall", "name": "ingressnodefirewallconfig", "reconcileID": "75076a96-84cf-49a4-ac58-3be86f8f0a96", "error": "could not update status for object &{TypeMeta:{Kind:IngressNodeFirewallConfig APIVersion:ingressnodefirewall.openshift.io/v1alpha1} ObjectMeta:{Name:ingressnodefirewallconfig GenerateName: Namespace:openshift-ingress-node-firewall SelfLink: UID:a31e4bc5-2262-41af-8262-8c0934be19e1 ResourceVersion:57963 Generation:1 CreationTimestamp:2022-10-27 21:52:12 +0000 UTC DeletionTimestamp:<nil> DeletionGracePeriodSeconds:<nil> Labels:map[] Annotations:map[] OwnerReferences:[] Finalizers:[] ManagedFields:[{Manager:kubectl-create Operation:Update APIVersion:ingressnodefirewall.openshift.io/v1alpha1 Time:2022-10-27 21:52:12 +0000 UTC FieldsType:FieldsV1 FieldsV1:{\"f:spec\":{\".\":{},\"f:nodeSelector\":{\".\":{},\"f:node-role.kubernetes.io/worker\":{}}}} Subresource:} {Manager:manager Operation:Update APIVersion:ingressnodefirewall.openshift.io/v1alpha1 Time:2022-10-27 21:55:44 +0000 UTC FieldsType:FieldsV1 FieldsV1:{\"f:status\":{\".\":{},\"f:conditions\":{}}} Subresource:status}]} Spec:{NodeSelector:map[node-role.kubernetes.io/worker:]} Status:{Conditions:[{Type:Available Status:True ObservedGeneration:0 LastTransitionTime:2022-10-27 22:00:10.290982302 +0000 UTC m=+20.432947581 Reason:Available Message:} {Type:Progressing Status:False ObservedGeneration:0 LastTransitionTime:2022-10-27 22:00:10.290982302 +0000 UTC m=+20.432947581 Reason:Progressing Message:} {Type:Degraded Status:False ObservedGeneration:0 LastTransitionTime:2022-10-27 22:00:10.290982302 +0000 UTC m=+20.432947581 Reason:Degraded Message:}]}}: Operation cannot be fulfilled on ingressnodefirewallconfigs.ingressnodefirewall.openshift.io \"ingressnodefirewallconfig\": the object has been modified; please apply your changes to the latest version and try again", "errorVerbose": "Operation cannot be fulfilled on ingressnodefirewallconfigs.ingressnodefirewall.openshift.io \"ingressnodefirewallconfig\": the object has been modified; please apply your changes to the latest version and try again\ncould not update status for object &{TypeMeta:{Kind:IngressNodeFirewallConfig APIVersion:ingressnodefirewall.openshift.io/v1alpha1} ObjectMeta:{Name:ingressnodefirewallconfig GenerateName: Namespace:openshift-ingress-node-firewall SelfLink: UID:a31e4bc5-2262-41af-8262-8c0934be19e1 ResourceVersion:57963 Generation:1 CreationTimestamp:2022-10-27 21:52:12 +0000 UTC DeletionTimestamp:<nil> DeletionGracePeriodSeconds:<nil> Labels:map[] Annotations:map[] OwnerReferences:[] Finalizers:[] ManagedFields:[{Manager:kubectl-create Operation:Update APIVersion:ingressnodefirewall.openshift.io/v1alpha1 Time:2022-10-27 21:52:12 +0000 UTC FieldsType:FieldsV1 FieldsV1:{\"f:spec\":{\".\":{},\"f:nodeSelector\":{\".\":{},\"f:node-role.kubernetes.io/worker\":{}}}} Subresource:} {Manager:manager Operation:Update APIVersion:ingressnodefirewall.openshift.io/v1alpha1 Time:2022-10-27 21:55:44 +0000 UTC FieldsType:FieldsV1 FieldsV1:{\"f:status\":{\".\":{},\"f:conditions\":{}}} Subresource:status}]} Spec:{NodeSelector:map[node-role.kubernetes.io/worker:]} Status:{Conditions:[{Type:Available Status:True ObservedGeneration:0 LastTransitionTime:2022-10-27 22:00:10.290982302 +0000 UTC m=+20.432947581 Reason:Available Message:} {Type:Progressing Status:False ObservedGeneration:0 LastTransitionTime:2022-10-27 22:00:10.290982302 +0000 UTC m=+20.432947581 Reason:Progressing Message:} {Type:Degraded Status:False ObservedGeneration:0 LastTransitionTime:2022-10-27 22:00:10.290982302 +0000 UTC m=+20.432947581 Reason:Degraded Message:}]}}\ngithub.com/openshift/ingress-node-firewall/pkg/status.Update\n\t/workspace/pkg/status/status.go:49\ngithub.com/openshift/ingress-node-firewall/controllers.(*IngressNodeFirewallConfigReconciler).Reconcile\n\t/workspace/controllers/ingressnodefirewallconfig_controller.go:115\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Reconcile\n\t/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:121\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler\n\t/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:320\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem\n\t/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:273\nsigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2\n\t/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:234\nruntime.goexit\n\t/usr/lib/golang/src/runtime/asm_amd64.s:1594"}
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler
/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:326
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem
/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:273
sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2
/workspace/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:234
1.6669080103081396e+09 INFO controllers.IngressNodeFirewallConfig.syncIngressNodeFirewallConfigResources Start
2022/10/27 22:00:10 reconciling (apps/v1, Kind=DaemonSet) openshift-ingress-node-firewall/ingress-node-firewall-daemon
2022/10/27 22:00:10 update was successful
1.666908010337311e+09 INFO controllers.IngressNodeFirewallConfig.syncIngressNodeFirewallConfigResources Start
2022/10/27 22:00:10 reconciling (apps/v1, Kind=DaemonSet) openshift-ingress-node-firewall/ingress-node-firewall-daemon
2022/10/27 22:00:10 update was successful
1.6669081197691467e+09 INFO controllers.IngressNodeFirewallConfig.syncIngressNodeFirewallConfigResources Start
2022/10/27 22:01:59 reconciling (apps/v1, Kind=DaemonSet) openshift-ingress-node-firewall/ingress-node-firewall-daemon
2022/10/27 22:01:59 update was successful
1.6669081197933285e+09 INFO controllers.IngressNodeFirewallConfig.syncIngressNodeFirewallConfigResources Start
2022/10/27 22:01:59 reconciling (apps/v1, Kind=DaemonSet) openshift-ingress-node-firewall/ingress-node-firewall-daemon
2022/10/27 22:01:59 update was successful
- blocks
-
OCPBUGS-3019 Ingress node firewall pod 's events container on the node causing pod in CrashLoopBackOff state when sctp module is loaded on node
-
- Closed
-
- is cloned by
-
OCPBUGS-3019 Ingress node firewall pod 's events container on the node causing pod in CrashLoopBackOff state when sctp module is loaded on node
-
- Closed
-
- links to