-
Bug
-
Resolution: Duplicate
-
Undefined
-
None
-
4.11
-
None
-
Quality / Stability / Reliability
-
False
-
-
None
-
Important
-
No
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
-
None
Description of problem:
During the upgrade mutlijob for OCP starting from version 4.10 with OVNkubernetes network type on OSP 16.2, the upgrade process encountered an error when upgrading from version 4.11 to 4.12. The operator 'network' is degraded.
Version-Release number of selected component (if applicable):
4.11.55 to 4.12.46 OVNkubernetes network RHOS-16.2-RHEL-8-20230510.n.1
How reproducible:
Always
Steps to Reproduce:
1.Begin the OCP upgrade process starting from version 4.10 2.Upgrade from 4.10 to 4.11 3.Upgrade from 4.11 to 4.12
Actual results:
'network' operator is degraded.
Expected results:
Smooth upgrade from 4.11 to 4.12 without any issues.
Additional info:
$ oc get co
NAME VERSION AVAILABLE PROGRESSING DEGRADED SINCE MESSAGE
authentication 4.12.46 True False False 14h
baremetal 4.12.46 True False False 16h
cloud-controller-manager 4.12.46 True False False 16h
cloud-credential 4.12.46 True False False 16h
cluster-autoscaler 4.12.46 True False False 16h
config-operator 4.12.46 True False False 16h
console 4.12.46 True False False 14h
control-plane-machine-set 4.12.46 True False False 13h
csi-snapshot-controller 4.12.46 True False False 16h
dns 4.11.55 True False False 16h
etcd 4.12.46 True False False 16h
image-registry 4.12.46 True False False 13h
ingress 4.12.46 True False False 15h
insights 4.12.46 True False False 16h
kube-apiserver 4.12.46 True False False 16h
kube-controller-manager 4.12.46 True False False 16h
kube-scheduler 4.12.46 True False False 16h
kube-storage-version-migrator 4.12.46 True False False 14h
machine-api 4.12.46 True False False 16h
machine-approver 4.12.46 True False False 16h
machine-config 4.11.55 True False False 14h
marketplace 4.12.46 True False False 16h
monitoring 4.12.46 True False False 15h
network 4.11.55 True False True 16h Error while updating operator configuration: could not apply (apps/v1, Kind=DaemonSet) openshift-ovn-kubernetes/ovnkube-master: failed to apply / update (apps/v1, Kind=DaemonSet) openshift-ovn-kubernetes/ovnkube-master: DaemonSet.apps "ovnkube-master" is invalid: [spec.template.spec.containers[1].lifecycle.preStop: Required value: must specify a handler type, spec.template.spec.containers[3].lifecycle.preStop: Required value: must specify a handler type]
node-tuning 4.12.46 True False False 13h
openshift-apiserver 4.12.46 True False False 16h
openshift-controller-manager 4.12.46 True False False 13h
openshift-samples 4.12.46 True False False 13h
operator-lifecycle-manager 4.12.46 True False False 16h
operator-lifecycle-manager-catalog 4.12.46 True False False 16h
operator-lifecycle-manager-packageserver 4.12.46 True False False 16h
service-ca 4.12.46 True False False 16h
storage 4.12.46 True False False 16h
$ oc get pods -n openshift-ovn-kubernetes
NAME READY STATUS RESTARTS AGE
ovnkube-master-9p9g4 6/6 Running 6 14h
ovnkube-master-x4hmq 6/6 Running 6 14h
ovnkube-master-xgbch 6/6 Running 6 14h
ovnkube-node-25jjs 5/5 Running 0 12h
ovnkube-node-b2qw8 5/5 Running 0 12h
ovnkube-node-ckw65 5/5 Running 3 (12h ago) 12h
ovnkube-node-gqssc 5/5 Running 0 12h
ovnkube-node-l4dcl 5/5 Running 0 12h
ovnkube-node-l5b7p 5/5 Running 0 12h
$ oc logs daemonset/ovnkube-master -n openshift-ovn-kubernetes
Found 3 pods, using pod/ovnkube-master-x4hmq
Defaulted container "northd" out of: northd, nbdb, kube-rbac-proxy, sbdb, ovnkube-master, ovn-dbchecker
+ [[ -f /env/_master ]]
+ trap quit TERM INT
++ date -Iseconds
+ echo '2023-12-24T16:38:59+00:00 - starting ovn-northd'
2023-12-24T16:38:59+00:00 - starting ovn-northd
+ wait 8
+ exec ovn-northd --no-chdir -vconsole:info -vfile:off '-vPATTERN:console:%D{%Y-%m-%dT%H:%M:%S.###Z}|%05N|%c%T|%p|%m' --ovnnb-db ssl:10.196.2.199:9641,ssl:10.196.2.200:9641,ssl:10.196.2.27:9641 --ovnsb-db ssl:10.196.2.199:9642,ssl:10.196.2.200:9642,ssl:10.196.2.27:9642 --pidfile /var/run/ovn/ovn-northd.pid --n-threads=4 -p /ovn-cert/tls.key -c /ovn-cert/tls.crt -C /ovn-ca/ca-bundle.crt
2023-12-24T16:39:00.103Z|00001|ovn_northd|INFO|Using 4 threads
2023-12-24T16:39:00.103Z|00002|ovn_northd|INFO|OVN internal version is : [22.12.3-20.27.0-70.6]
2023-12-24T16:39:00.104Z|00003|ovn_parallel_hmap|INFO|Setting thread count to 4
2023-12-24T16:39:00.104Z|00004|ovn_parallel_hmap|INFO|Creating new pool with size 4
2023-12-24T16:39:00.111Z|00005|reconnect|INFO|ssl:10.196.2.199:9641: connecting...
2023-12-24T16:39:00.111Z|00006|reconnect|INFO|ssl:10.196.2.199:9641: connection attempt failed (Connection refused)
2023-12-24T16:39:00.111Z|00007|reconnect|INFO|ssl:10.196.2.27:9641: connecting...
2023-12-24T16:39:00.111Z|00008|ovn_northd|INFO|OVN NB IDL reconnected, force recompute.
2023-12-24T16:39:00.111Z|00009|reconnect|INFO|ssl:10.196.2.27:9642: connecting...
2023-12-24T16:39:00.111Z|00010|ovn_northd|INFO|OVN SB IDL reconnected, force recompute.
2023-12-24T16:39:00.120Z|00011|reconnect|INFO|ssl:10.196.2.27:9641: connected
2023-12-24T16:39:00.124Z|00012|reconnect|INFO|ssl:10.196.2.27:9642: connected
2023-12-24T16:39:10.104Z|00013|memory|INFO|24216 kB peak resident set size after 10.0 seconds
2023-12-24T16:50:47.153Z|00014|stream_ssl|WARN|SSL_read: unexpected SSL connection close
2023-12-24T16:50:47.153Z|00015|jsonrpc|WARN|ssl:10.196.2.27:9642: receive error: Protocol error
2023-12-24T16:50:47.154Z|00016|reconnect|WARN|ssl:10.196.2.27:9642: connection dropped (Protocol error)
2023-12-24T16:50:47.154Z|00017|reconnect|INFO|ssl:10.196.2.199:9642: connecting...
2023-12-24T16:50:47.158Z|00018|reconnect|INFO|ssl:10.196.2.199:9642: connected
2023-12-24T16:50:47.161Z|00019|ovn_northd|INFO|ovn-northd lock acquired. This ovn-northd instance is now active.
2023-12-24T16:50:47.162Z|00020|ovsdb_cs|INFO|ssl:10.196.2.199:9642: clustered database server is not cluster leader; trying another server
2023-12-24T16:50:47.162Z|00021|reconnect|INFO|ssl:10.196.2.199:9642: connection attempt timed out
2023-12-24T16:50:47.162Z|00022|ovn_northd|INFO|ovn-northd lock lost. This ovn-northd instance is now on standby.
2023-12-24T16:50:47.162Z|00023|reconnect|INFO|ssl:10.196.2.200:9642: connecting...
2023-12-24T16:50:47.166Z|00024|reconnect|INFO|ssl:10.196.2.200:9642: connected
2023-12-24T16:50:47.167Z|00025|ovn_northd|INFO|ovn-northd lock acquired. This ovn-northd instance is now active.
2023-12-24T16:50:47.168Z|00026|ovsdb_cs|INFO|ssl:10.196.2.200:9642: clustered database server is not cluster leader; trying another server
2023-12-24T16:50:47.168Z|00027|reconnect|INFO|ssl:10.196.2.200:9642: connection attempt timed out
2023-12-24T16:50:47.169Z|00028|ovn_northd|INFO|ovn-northd lock lost. This ovn-northd instance is now on standby.
2023-12-24T16:50:47.173Z|00029|stream_ssl|WARN|SSL_read: unexpected SSL connection close
2023-12-24T16:50:47.173Z|00030|jsonrpc|WARN|ssl:10.196.2.27:9641: receive error: Protocol error
2023-12-24T16:50:47.173Z|00031|reconnect|WARN|ssl:10.196.2.27:9641: connection dropped (Protocol error)
2023-12-24T16:50:47.173Z|00032|reconnect|INFO|ssl:10.196.2.200:9641: connecting...
2023-12-24T16:50:47.184Z|00033|reconnect|INFO|ssl:10.196.2.200:9641: connected
2023-12-24T16:50:47.187Z|00034|ovsdb_cs|INFO|ssl:10.196.2.200:9641: clustered database server is not cluster leader; trying another server
2023-12-24T16:50:47.187Z|00035|reconnect|INFO|ssl:10.196.2.200:9641: connection attempt timed out
2023-12-24T16:50:48.169Z|00036|reconnect|INFO|ssl:10.196.2.27:9642: connecting...
2023-12-24T16:50:48.169Z|00037|reconnect|INFO|ssl:10.196.2.27:9642: connection attempt failed (Connection refused)
2023-12-24T16:50:48.170Z|00038|reconnect|INFO|ssl:10.196.2.27:9642: waiting 2 seconds before reconnect
2023-12-24T16:50:48.188Z|00039|reconnect|INFO|ssl:10.196.2.199:9641: connecting...
2023-12-24T16:50:48.201Z|00040|reconnect|INFO|ssl:10.196.2.199:9641: connected
2023-12-24T16:50:48.205Z|00041|ovsdb_cs|INFO|ssl:10.196.2.199:9641: clustered database server is not cluster leader; trying another server
2023-12-24T16:50:48.205Z|00042|reconnect|INFO|ssl:10.196.2.199:9641: connection attempt timed out
2023-12-24T16:50:48.205Z|00043|reconnect|INFO|ssl:10.196.2.199:9641: waiting 2 seconds before reconnect
2023-12-24T16:50:50.171Z|00044|reconnect|INFO|ssl:10.196.2.199:9642: connecting...
2023-12-24T16:50:50.176Z|00045|reconnect|INFO|ssl:10.196.2.199:9642: connected
2023-12-24T16:50:50.179Z|00046|ovn_northd|INFO|ovn-northd lock acquired. This ovn-northd instance is now active.
2023-12-24T16:50:50.180Z|00047|ovsdb_cs|INFO|ssl:10.196.2.199:9642: clustered database server is not cluster leader; trying another server
2023-12-24T16:50:50.180Z|00048|reconnect|INFO|ssl:10.196.2.199:9642: connection attempt timed out
2023-12-24T16:50:50.181Z|00049|reconnect|INFO|ssl:10.196.2.199:9642: waiting 4 seconds before reconnect
2023-12-24T16:50:50.181Z|00050|ovn_northd|INFO|ovn-northd lock lost. This ovn-northd instance is now on standby.
2023-12-24T16:50:50.208Z|00051|reconnect|INFO|ssl:10.196.2.27:9641: connecting...
2023-12-24T16:50:50.208Z|00052|reconnect|INFO|ssl:10.196.2.27:9641: connection attempt failed (Connection refused)
2023-12-24T16:50:50.208Z|00053|reconnect|INFO|ssl:10.196.2.27:9641: waiting 4 seconds before reconnect
2023-12-24T16:50:54.184Z|00054|reconnect|INFO|ssl:10.196.2.200:9642: connecting...
2023-12-24T16:50:54.190Z|00055|reconnect|INFO|ssl:10.196.2.200:9642: connected
2023-12-24T16:50:54.191Z|00056|ovn_northd|INFO|ovn-northd lock acquired. This ovn-northd instance is now active.
2023-12-24T16:50:54.192Z|00057|ovsdb_cs|INFO|ssl:10.196.2.200:9642: clustered database server is not cluster leader; trying another server
2023-12-24T16:50:54.192Z|00058|reconnect|INFO|ssl:10.196.2.200:9642: connection attempt timed out
2023-12-24T16:50:54.192Z|00059|reconnect|INFO|ssl:10.196.2.200:9642: continuing to reconnect in the background but suppressing further logging
2023-12-24T16:50:54.192Z|00060|ovn_northd|INFO|ovn-northd lock lost. This ovn-northd instance is now on standby.
2023-12-24T16:50:54.209Z|00061|reconnect|INFO|ssl:10.196.2.200:9641: connecting...
2023-12-24T16:50:54.212Z|00062|reconnect|INFO|ssl:10.196.2.200:9641: connected
2023-12-24T16:50:54.216Z|00063|ovsdb_cs|INFO|ssl:10.196.2.200:9641: clustered database server is not cluster leader; trying another server
2023-12-24T16:50:54.216Z|00064|reconnect|INFO|ssl:10.196.2.200:9641: connection attempt timed out
2023-12-24T16:50:54.216Z|00065|reconnect|INFO|ssl:10.196.2.200:9641: continuing to reconnect in the background but suppressing further logging
2023-12-24T16:51:02.224Z|00066|reconnect|INFO|ssl:10.196.2.199:9641: connected
2023-12-24T16:51:02.226Z|00067|ovsdb_cs|INFO|ssl:10.196.2.199:9641: clustered database server is not cluster leader; trying another server
2023-12-24T16:51:10.218Z|00068|reconnect|INFO|ssl:10.196.2.199:9642: connected
2023-12-24T16:51:18.252Z|00069|reconnect|INFO|ssl:10.196.2.200:9641: connected
2023-12-24T16:58:32.608Z|00070|ovsdb_cs|INFO|ssl:10.196.2.199:9642: clustered database server is not cluster leader; trying another server
2023-12-24T16:58:32.609Z|00071|reconnect|INFO|ssl:10.196.2.200:9642: connecting...
2023-12-24T16:58:32.622Z|00072|reconnect|INFO|ssl:10.196.2.200:9642: connected
2023-12-24T16:58:32.627Z|00073|ovsdb_cs|INFO|ssl:10.196.2.200:9642: clustered database server is not cluster leader; trying another server
2023-12-24T16:58:32.627Z|00074|reconnect|INFO|ssl:10.196.2.200:9642: connection attempt timed out
2023-12-24T16:58:32.630Z|00075|reconnect|INFO|ssl:10.196.2.27:9642: connecting...
2023-12-24T16:58:32.635Z|00076|reconnect|INFO|ssl:10.196.2.27:9642: connected
- duplicates
-
OCPBUGS-24039 [4.12] CNO fails to apply ovnkube-master daemonset during upgrade
-
- Closed
-