-
Bug
-
Resolution: Duplicate
-
Major
-
None
-
4.13, 4.12, 4.14
-
None
-
Quality / Stability / Reliability
-
False
-
-
None
-
Important
-
Yes
-
None
-
None
-
Rejected
-
CLOUD Sprint 239
-
1
-
None
-
None
-
None
-
None
-
None
-
None
-
None
Description of problem:
MHC for control plane cannot work right for some cases Tried three cases: 1.Terminate/Delete a master on the cloud provider console, no issue 2.Stop the kubelet service on the master node, new master get Running, the old one stuck in Deleting, many co degraded. 3.Delete a master node, the old machine stuck in Deleting, no new machine created. This is a regression bug, because I tested this on 4.12 around September 2022, case 2 and case 3 work right. https://polarion.engineering.redhat.com/polarion/#/project/OSE/workitem?id=OCP-54326
Version-Release number of selected component (if applicable):
4.14.0-0.nightly-2023-06-05-112833 4.13.0-0.nightly-2023-06-06-194351 4.12.0-0.nightly-2023-06-07-005319
How reproducible:
Always
Steps to Reproduce:
1.Create MHC for control plane
apiVersion: machine.openshift.io/v1beta1
kind: MachineHealthCheck
metadata:
name: control-plane-health
namespace: openshift-machine-api
spec:
maxUnhealthy: 1
selector:
matchLabels:
machine.openshift.io/cluster-api-machine-type: master
unhealthyConditions:
- status: "False"
timeout: 300s
type: Ready
- status: "Unknown"
timeout: 300s
type: Ready
liuhuali@Lius-MacBook-Pro huali-test % oc create -f mhc-master3.yaml
machinehealthcheck.machine.openshift.io/control-plane-health created
liuhuali@Lius-MacBook-Pro huali-test % oc get mhc
NAME MAXUNHEALTHY EXPECTEDMACHINES CURRENTHEALTHY
control-plane-health 1 3 3
machine-api-termination-handler 100% 0 0
Case 1: Terminate/Delete a master on the cloud provider console, no issue.
liuhuali@Lius-MacBook-Pro huali-test % oc get node
NAME STATUS ROLES AGE VERSION
huliu-az7c-svq9q-master-0 Ready control-plane,master 70m v1.26.5+7a891f0
huliu-az7c-svq9q-master-1 Ready control-plane,master 69m v1.26.5+7a891f0
huliu-az7c-svq9q-master-2 Ready control-plane,master 69m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-5r8jf Ready worker 8m48s v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-k747l Ready worker 21m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-r2vdn Ready worker 57m v1.26.5+7a891f0
liuhuali@Lius-MacBook-Pro huali-test % oc get node
NAME STATUS ROLES AGE VERSION
huliu-az7c-svq9q-master-1 Ready control-plane,master 70m v1.26.5+7a891f0
huliu-az7c-svq9q-master-2 Ready control-plane,master 70m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-5r8jf Ready worker 9m28s v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-k747l Ready worker 22m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-r2vdn Ready worker 58m v1.26.5+7a891f0
liuhuali@Lius-MacBook-Pro huali-test % oc get machine
NAME PHASE TYPE REGION ZONE AGE
huliu-az7c-svq9q-master-0 Deleting Standard_D8s_v3 westus 72m
huliu-az7c-svq9q-master-1 Running Standard_D8s_v3 westus 72m
huliu-az7c-svq9q-master-2 Running Standard_D8s_v3 westus 72m
huliu-az7c-svq9q-worker-westus-5r8jf Running Standard_D4s_v3 westus 15m
huliu-az7c-svq9q-worker-westus-k747l Running Standard_D4s_v3 westus 28m
huliu-az7c-svq9q-worker-westus-r2vdn Running Standard_D4s_v3 westus 66m
liuhuali@Lius-MacBook-Pro huali-test % oc get machine
NAME PHASE TYPE REGION ZONE AGE
huliu-az7c-svq9q-master-1 Running Standard_D8s_v3 westus 74m
huliu-az7c-svq9q-master-2 Running Standard_D8s_v3 westus 74m
huliu-az7c-svq9q-master-c96k8-0 Provisioning 2s
huliu-az7c-svq9q-worker-westus-5r8jf Running Standard_D4s_v3 westus 16m
huliu-az7c-svq9q-worker-westus-k747l Running Standard_D4s_v3 westus 29m
huliu-az7c-svq9q-worker-westus-r2vdn Running Standard_D4s_v3 westus 67m
liuhuali@Lius-MacBook-Pro huali-test % oc get machine
NAME PHASE TYPE REGION ZONE AGE
huliu-az7c-svq9q-master-1 Running Standard_D8s_v3 westus 95m
huliu-az7c-svq9q-master-2 Running Standard_D8s_v3 westus 95m
huliu-az7c-svq9q-master-c96k8-0 Running Standard_D8s_v3 westus 21m
huliu-az7c-svq9q-worker-westus-5r8jf Running Standard_D4s_v3 westus 37m
huliu-az7c-svq9q-worker-westus-k747l Running Standard_D4s_v3 westus 50m
huliu-az7c-svq9q-worker-westus-r2vdn Running Standard_D4s_v3 westus 88m
liuhuali@Lius-MacBook-Pro huali-test % oc get node
NAME STATUS ROLES AGE VERSION
huliu-az7c-svq9q-master-1 Ready control-plane,master 93m v1.26.5+7a891f0
huliu-az7c-svq9q-master-2 Ready control-plane,master 93m v1.26.5+7a891f0
huliu-az7c-svq9q-master-c96k8-0 Ready control-plane,master 17m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-5r8jf Ready worker 32m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-k747l Ready worker 44m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-r2vdn Ready worker 80m v1.26.5+7a891f0
Case 2.Stop the kubelet service on the master node, new master get Running, the old one stuck in Deleting, many co degraded.
liuhuali@Lius-MacBook-Pro huali-test % oc debug node/huliu-az7c-svq9q-master-1
Starting pod/huliu-az7c-svq9q-master-1-debug ...
To use host binaries, run `chroot /host`
Pod IP: 10.0.0.6
If you don't see a command prompt, try pressing enter.
sh-4.4# chroot /host
sh-5.1# systemctl stop kubelet
Removing debug pod ...
liuhuali@Lius-MacBook-Pro huali-test % oc get node
NAME STATUS ROLES AGE VERSION
huliu-az7c-svq9q-master-1 Ready control-plane,master 95m v1.26.5+7a891f0
huliu-az7c-svq9q-master-2 Ready control-plane,master 95m v1.26.5+7a891f0
huliu-az7c-svq9q-master-c96k8-0 Ready control-plane,master 19m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-5r8jf Ready worker 34m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-k747l Ready worker 47m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-r2vdn Ready worker 83m v1.26.5+7a891f0
liuhuali@Lius-MacBook-Pro huali-test % oc get machine
NAME PHASE TYPE REGION ZONE AGE
huliu-az7c-svq9q-master-1 Running Standard_D8s_v3 westus 97m
huliu-az7c-svq9q-master-2 Running Standard_D8s_v3 westus 97m
huliu-az7c-svq9q-master-c96k8-0 Running Standard_D8s_v3 westus 23m
huliu-az7c-svq9q-worker-westus-5r8jf Running Standard_D4s_v3 westus 39m
huliu-az7c-svq9q-worker-westus-k747l Running Standard_D4s_v3 westus 53m
huliu-az7c-svq9q-worker-westus-r2vdn Running Standard_D4s_v3 westus 91m
liuhuali@Lius-MacBook-Pro huali-test % oc get node
NAME STATUS ROLES AGE VERSION
huliu-az7c-svq9q-master-1 NotReady control-plane,master 107m v1.26.5+7a891f0
huliu-az7c-svq9q-master-2 Ready control-plane,master 107m v1.26.5+7a891f0
huliu-az7c-svq9q-master-c96k8-0 Ready control-plane,master 32m v1.26.5+7a891f0
huliu-az7c-svq9q-master-jdhgg-1 Ready control-plane,master 2m10s v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-5r8jf Ready worker 46m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-k747l Ready worker 59m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-r2vdn Ready worker 95m v1.26.5+7a891f0
liuhuali@Lius-MacBook-Pro huali-test % oc get machine
NAME PHASE TYPE REGION ZONE AGE
huliu-az7c-svq9q-master-1 Deleting Standard_D8s_v3 westus 110m
huliu-az7c-svq9q-master-2 Running Standard_D8s_v3 westus 110m
huliu-az7c-svq9q-master-c96k8-0 Running Standard_D8s_v3 westus 36m
huliu-az7c-svq9q-master-jdhgg-1 Running Standard_D8s_v3 westus 5m55s
huliu-az7c-svq9q-worker-westus-5r8jf Running Standard_D4s_v3 westus 52m
huliu-az7c-svq9q-worker-westus-k747l Running Standard_D4s_v3 westus 65m
huliu-az7c-svq9q-worker-westus-r2vdn Running Standard_D4s_v3 westus 103m
liuhuali@Lius-MacBook-Pro huali-test % oc get machine
NAME PHASE TYPE REGION ZONE AGE
huliu-az7c-svq9q-master-1 Deleting Standard_D8s_v3 westus 3h
huliu-az7c-svq9q-master-2 Running Standard_D8s_v3 westus 3h
huliu-az7c-svq9q-master-c96k8-0 Running Standard_D8s_v3 westus 105m
huliu-az7c-svq9q-master-jdhgg-1 Running Standard_D8s_v3 westus 75m
huliu-az7c-svq9q-worker-westus-5r8jf Running Standard_D4s_v3 westus 122m
huliu-az7c-svq9q-worker-westus-k747l Running Standard_D4s_v3 westus 135m
huliu-az7c-svq9q-worker-westus-r2vdn Running Standard_D4s_v3 westus 173m
liuhuali@Lius-MacBook-Pro huali-test % oc get node
NAME STATUS ROLES AGE VERSION
huliu-az7c-svq9q-master-1 NotReady control-plane,master 178m v1.26.5+7a891f0
huliu-az7c-svq9q-master-2 Ready control-plane,master 178m v1.26.5+7a891f0
huliu-az7c-svq9q-master-c96k8-0 Ready control-plane,master 102m v1.26.5+7a891f0
huliu-az7c-svq9q-master-jdhgg-1 Ready control-plane,master 72m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-5r8jf Ready worker 116m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-k747l Ready worker 129m v1.26.5+7a891f0
huliu-az7c-svq9q-worker-westus-r2vdn Ready worker 165m v1.26.5+7a891f0
liuhuali@Lius-MacBook-Pro huali-test % oc get co
NAME VERSION AVAILABLE PROGRESSING DEGRADED SINCE MESSAGE
authentication 4.13.0-0.nightly-2023-06-06-194351 True True True 107m APIServerDeploymentDegraded: 1 of 4 requested instances are unavailable for apiserver.openshift-oauth-apiserver ()...
baremetal 4.13.0-0.nightly-2023-06-06-194351 True False False 174m
cloud-controller-manager 4.13.0-0.nightly-2023-06-06-194351 True False False 176m
cloud-credential 4.13.0-0.nightly-2023-06-06-194351 True False False 3h
cluster-autoscaler 4.13.0-0.nightly-2023-06-06-194351 True False False 173m
config-operator 4.13.0-0.nightly-2023-06-06-194351 True False False 175m
console 4.13.0-0.nightly-2023-06-06-194351 True False False 136m
control-plane-machine-set 4.13.0-0.nightly-2023-06-06-194351 True False False 71m
csi-snapshot-controller 4.13.0-0.nightly-2023-06-06-194351 True False False 174m
dns 4.13.0-0.nightly-2023-06-06-194351 True True False 173m DNS "default" reports Progressing=True: "Have 6 available node-resolver pods, want 7."
etcd 4.13.0-0.nightly-2023-06-06-194351 True True True 173m NodeControllerDegraded: The master nodes not ready: node "huliu-az7c-svq9q-master-1" not ready since 2023-06-07 08:47:34 +0000 UTC because NodeStatusUnknown (Kubelet stopped posting node status.)
image-registry 4.13.0-0.nightly-2023-06-06-194351 True True False 165m Progressing: The registry is ready...
ingress 4.13.0-0.nightly-2023-06-06-194351 True False False 165m
insights 4.13.0-0.nightly-2023-06-06-194351 True False False 168m
kube-apiserver 4.13.0-0.nightly-2023-06-06-194351 True True True 171m NodeControllerDegraded: The master nodes not ready: node "huliu-az7c-svq9q-master-1" not ready since 2023-06-07 08:47:34 +0000 UTC because NodeStatusUnknown (Kubelet stopped posting node status.)
kube-controller-manager 4.13.0-0.nightly-2023-06-06-194351 True False True 171m NodeControllerDegraded: The master nodes not ready: node "huliu-az7c-svq9q-master-1" not ready since 2023-06-07 08:47:34 +0000 UTC because NodeStatusUnknown (Kubelet stopped posting node status.)
kube-scheduler 4.13.0-0.nightly-2023-06-06-194351 True False True 171m NodeControllerDegraded: The master nodes not ready: node "huliu-az7c-svq9q-master-1" not ready since 2023-06-07 08:47:34 +0000 UTC because NodeStatusUnknown (Kubelet stopped posting node status.)
kube-storage-version-migrator 4.13.0-0.nightly-2023-06-06-194351 True False False 106m
machine-api 4.13.0-0.nightly-2023-06-06-194351 True False False 167m
machine-approver 4.13.0-0.nightly-2023-06-06-194351 True False False 174m
machine-config 4.13.0-0.nightly-2023-06-06-194351 False False True 60m Cluster not available for [{operator 4.13.0-0.nightly-2023-06-06-194351}]: failed to apply machine config daemon manifests: error during waitForDaemonsetRollout: [timed out waiting for the condition, daemonset machine-config-daemon is not ready. status: (desired: 7, updated: 7, ready: 6, unavailable: 1)]
marketplace 4.13.0-0.nightly-2023-06-06-194351 True False False 174m
monitoring 4.13.0-0.nightly-2023-06-06-194351 True False False 106m
network 4.13.0-0.nightly-2023-06-06-194351 True True False 177m DaemonSet "/openshift-network-diagnostics/network-check-target" is not available (awaiting 1 nodes)...
node-tuning 4.13.0-0.nightly-2023-06-06-194351 True False False 173m
openshift-apiserver 4.13.0-0.nightly-2023-06-06-194351 True True True 107m APIServerDeploymentDegraded: 1 of 4 requested instances are unavailable for apiserver.openshift-apiserver ()
openshift-controller-manager 4.13.0-0.nightly-2023-06-06-194351 True False False 170m
openshift-samples 4.13.0-0.nightly-2023-06-06-194351 True False False 167m
operator-lifecycle-manager 4.13.0-0.nightly-2023-06-06-194351 True False False 174m
operator-lifecycle-manager-catalog 4.13.0-0.nightly-2023-06-06-194351 True False False 174m
operator-lifecycle-manager-packageserver 4.13.0-0.nightly-2023-06-06-194351 True False False 168m
service-ca 4.13.0-0.nightly-2023-06-06-194351 True False False 175m
storage 4.13.0-0.nightly-2023-06-06-194351 True True False 174m AzureDiskCSIDriverOperatorCRProgressing: AzureDiskDriverNodeServiceControllerProgressing: Waiting for DaemonSet to deploy node pods...
liuhuali@Lius-MacBook-Pro huali-test %
Case 3.Delete a master node, the old machine stuck in Deleting, no new machine created.
liuhuali@Lius-MacBook-Pro huali-test % oc delete node huliu-az7a-d2lgf-master-0
node "huliu-az7a-d2lgf-master-0" deleted
liuhuali@Lius-MacBook-Pro huali-test % oc get machine
NAME PHASE TYPE REGION ZONE AGE
huliu-az7a-d2lgf-master-0 Deleting Standard_D8s_v3 westus 4h58m
huliu-az7a-d2lgf-master-1 Running Standard_D8s_v3 westus 4h58m
huliu-az7a-d2lgf-master-rjxg9-2 Running Standard_D8s_v3 westus 27m
huliu-az7a-d2lgf-worker-westus-blbrj Running Standard_D4s_v3 westus 4h49m
huliu-az7a-d2lgf-worker-westus-r7g5l Running Standard_D4s_v3 westus 4h15m
huliu-az7a-d2lgf-worker-westus-rrc97 Running Standard_D4s_v3 westus 4h49m
liuhuali@Lius-MacBook-Pro huali-test % oc get node
NAME STATUS ROLES AGE VERSION
huliu-az7a-d2lgf-master-1 Ready control-plane,master 4h53m v1.27.2+cc041e8
huliu-az7a-d2lgf-master-rjxg9-2 Ready control-plane,master 23m v1.27.2+cc041e8
huliu-az7a-d2lgf-worker-westus-blbrj Ready worker 4h38m v1.27.2+cc041e8
huliu-az7a-d2lgf-worker-westus-r7g5l Ready worker 4h8m v1.27.2+cc041e8
huliu-az7a-d2lgf-worker-westus-rrc97 Ready worker 4h38m v1.27.2+cc041e8
liuhuali@Lius-MacBook-Pro huali-test % oc get machine
NAME PHASE TYPE REGION ZONE AGE
huliu-az7a-d2lgf-master-0 Deleting Standard_D8s_v3 westus 8h
huliu-az7a-d2lgf-master-1 Running Standard_D8s_v3 westus 8h
huliu-az7a-d2lgf-master-rjxg9-2 Running Standard_D8s_v3 westus 4h27m
huliu-az7a-d2lgf-worker-westus-8wnmq Running Standard_D4s_v3 westus 3h20m
huliu-az7a-d2lgf-worker-westus-df6wj Running Standard_D4s_v3 westus 167m
huliu-az7a-d2lgf-worker-westus-r7g5l Running Standard_D4s_v3 westus 8h
liuhuali@Lius-MacBook-Pro huali-test % oc get node
NAME STATUS ROLES AGE VERSION
huliu-az7a-d2lgf-master-1 Ready control-plane,master 8h v1.27.2+cc041e8
huliu-az7a-d2lgf-master-rjxg9-2 Ready control-plane,master 4h23m v1.27.2+cc041e8
huliu-az7a-d2lgf-worker-westus-8wnmq Ready worker 3h15m v1.27.2+cc041e8
huliu-az7a-d2lgf-worker-westus-df6wj Ready worker 158m v1.27.2+cc041e8
huliu-az7a-d2lgf-worker-westus-r7g5l Ready worker 8h v1.27.2+cc041e8
liuhuali@Lius-MacBook-Pro huali-test % oc get co
NAME VERSION AVAILABLE PROGRESSING DEGRADED SINCE MESSAGE
authentication 4.14.0-0.nightly-2023-06-05-112833 True False False 45m
baremetal 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
cloud-controller-manager 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
cloud-credential 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
cluster-autoscaler 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
config-operator 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
console 4.14.0-0.nightly-2023-06-05-112833 True False False 72m
control-plane-machine-set 4.14.0-0.nightly-2023-06-05-112833 True False False 4h21m
csi-snapshot-controller 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
dns 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
etcd 4.14.0-0.nightly-2023-06-05-112833 True False True 8h EtcdEndpointsDegraded: EtcdEndpointsController can't evaluate whether quorum is safe: CheckSafeToScaleCluster 3 nodes are required, but only 2 are available
image-registry 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
ingress 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
insights 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
kube-apiserver 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
kube-controller-manager 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
kube-scheduler 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
kube-storage-version-migrator 4.14.0-0.nightly-2023-06-05-112833 True False False 4h30m
machine-api 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
machine-approver 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
machine-config 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
marketplace 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
monitoring 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
network 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
node-tuning 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
openshift-apiserver 4.14.0-0.nightly-2023-06-05-112833 True False False 32m
openshift-controller-manager 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
openshift-samples 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
operator-lifecycle-manager 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
operator-lifecycle-manager-catalog 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
operator-lifecycle-manager-packageserver 4.14.0-0.nightly-2023-06-05-112833 False True False 166m ClusterServiceVersion openshift-operator-lifecycle-manager/packageserver observed in phase Failed with reason: ComponentUnhealthy, message: apiServices not installed
service-ca 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
storage 4.14.0-0.nightly-2023-06-05-112833 True False False 8h
liuhuali@Lius-MacBook-Pro huali-test %
liuhuali@Lius-MacBook-Pro huali-test % oc logs machine-api-controllers-5c58f965fd-nwjxh -c machine-healthcheck-controller
…
I0607 10:13:57.348074 1 machinehealthcheck_controller.go:162] Reconciling openshift-machine-api/control-plane-health
I0607 10:13:57.348318 1 machinehealthcheck_controller.go:188] Reconciling openshift-machine-api/control-plane-health: finding targets
I0607 10:13:57.348601 1 machinehealthcheck_controller.go:465] Reconciling openshift-machine-api/control-plane-health/huliu-az7a-d2lgf-master-0/huliu-az7a-d2lgf-master-0: health checking
I0607 10:13:57.348615 1 machinehealthcheck_controller.go:465] Reconciling openshift-machine-api/control-plane-health/huliu-az7a-d2lgf-master-1/huliu-az7a-d2lgf-master-1: health checking
I0607 10:13:57.348624 1 machinehealthcheck_controller.go:465] Reconciling openshift-machine-api/control-plane-health/huliu-az7a-d2lgf-master-rjxg9-2/huliu-az7a-d2lgf-master-rjxg9-2: health checking
I0607 10:13:57.348636 1 machinehealthcheck_controller.go:251] Remediations are allowed for openshift-machine-api/control-plane-health: total targets: 3, max unhealthy: 1, unhealthy targets: 1
I0607 10:13:57.360972 1 machinehealthcheck_controller.go:287] Reconciling openshift-machine-api/control-plane-health/huliu-az7a-d2lgf-master-0/huliu-az7a-d2lgf-master-0: meet unhealthy criteria, triggers remediation
I0607 10:13:57.361004 1 machinehealthcheck_controller.go:628] openshift-machine-api/control-plane-health/huliu-az7a-d2lgf-master-0/huliu-az7a-d2lgf-master-0: start remediation logic
I0607 10:13:57.361060 1 machinehealthcheck_controller.go:279] Reconciling openshift-machine-api/control-plane-health: no more targets meet unhealthy criteria
liuhuali@Lius-MacBook-Pro huali-test % oc logs control-plane-machine-set-operator-7cb46c5fdd-4lgrc
…
I0607 10:14:48.262167 1 controller.go:157] "msg"="Reconciling control plane machine set" "controller"="controlplanemachineset" "name"="cluster" "namespace"="openshift-machine-api" "reconcileID"="26c85d60-742b-48e9-b162-a398b170eb4c"
I0607 10:14:48.332215 1 controller.go:210] "msg"="Finished reconciling control plane machine set" "controller"="controlplanemachineset" "name"="cluster" "namespace"="openshift-machine-api" "reconcileID"="26c85d60-742b-48e9-b162-a398b170eb4c"
E0607 10:14:48.332285 1 controller.go:329] "msg"="Reconciler error" "error"="error reconciling control plane machine set: error fetching machine info: could not generate machine info for machine huliu-az7a-d2lgf-master-0: error checking machine readiness: failed to get Node \"huliu-az7a-d2lgf-master-0\": Node \"huliu-az7a-d2lgf-master-0\" not found" "controller"="controlplanemachineset" "reconcileID"="26c85d60-742b-48e9-b162-a398b170eb4c"
Actual results:
MHC for control plane cannot work right for case2 and case3
Expected results:
MHC for control plane should work right
Additional info:
Must gather for case2: https://drive.google.com/file/d/1eRa4tt7Mr5hMD8PVCOUgCLFHFvmx2J3T/view?usp=sharing Must gather for case3: https://drive.google.com/file/d/1ZdBXhQCDJvc2zcZhBmZaWicykzsbL3Z_/view?usp=sharing
- is cloned by
-
OCPBUGS-17199 CEO prevents member deletion during revision rollout
-
- Closed
-